[ 497.634676] env[63531]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=63531) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 497.635054] env[63531]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=63531) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 497.635192] env[63531]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=63531) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 497.635446] env[63531]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 497.727172] env[63531]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=63531) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 497.736742] env[63531]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=63531) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 498.337733] env[63531]: INFO nova.virt.driver [None req-952a15e7-d619-459c-92e0-0b3cb65e5c7c None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 498.407253] env[63531]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 498.407404] env[63531]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 498.407514] env[63531]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=63531) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 501.498093] env[63531]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-666b1e60-61fc-4b0e-83e1-a42d96752488 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 501.513908] env[63531]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=63531) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 501.514034] env[63531]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-b7b26013-ff10-4191-b141-7a936588583d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 501.540177] env[63531]: INFO oslo_vmware.api [-] Successfully established new session; session ID is bbc41. [ 501.540326] env[63531]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.133s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 501.540814] env[63531]: INFO nova.virt.vmwareapi.driver [None req-952a15e7-d619-459c-92e0-0b3cb65e5c7c None None] VMware vCenter version: 7.0.3 [ 501.544272] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4503d13e-e309-423f-909d-e45e648e2088 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 501.560989] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52bbefea-bccb-4091-a683-eca437f83b47 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 501.567161] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-456076bb-7d44-434e-b205-f640e9879ea4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 501.573688] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31e4f602-662f-413c-93d0-078f6e9a217e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 501.586494] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44e59186-3524-4e59-aceb-54a777875b45 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 501.592156] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78aad60e-496c-4489-b0a0-2cb545318164 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 501.621409] env[63531]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-6e56ee19-90c7-4c45-8c29-c031f470c181 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 501.626341] env[63531]: DEBUG nova.virt.vmwareapi.driver [None req-952a15e7-d619-459c-92e0-0b3cb65e5c7c None None] Extension org.openstack.compute already exists. {{(pid=63531) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:225}} [ 501.628920] env[63531]: INFO nova.compute.provider_config [None req-952a15e7-d619-459c-92e0-0b3cb65e5c7c None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 502.132055] env[63531]: DEBUG nova.context [None req-952a15e7-d619-459c-92e0-0b3cb65e5c7c None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),82819688-61a4-4a0c-802c-fcfc75bd49b7(cell1) {{(pid=63531) load_cells /opt/stack/nova/nova/context.py:464}} [ 502.134255] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 502.134496] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 502.135182] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 502.135625] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] Acquiring lock "82819688-61a4-4a0c-802c-fcfc75bd49b7" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 502.135987] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] Lock "82819688-61a4-4a0c-802c-fcfc75bd49b7" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 502.137149] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] Lock "82819688-61a4-4a0c-802c-fcfc75bd49b7" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 502.157562] env[63531]: INFO dbcounter [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] Registered counter for database nova_cell0 [ 502.165623] env[63531]: INFO dbcounter [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] Registered counter for database nova_cell1 [ 502.169100] env[63531]: DEBUG oslo_db.sqlalchemy.engines [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=63531) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 502.169705] env[63531]: DEBUG oslo_db.sqlalchemy.engines [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=63531) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 502.174110] env[63531]: ERROR nova.db.main.api [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 502.174110] env[63531]: result = function(*args, **kwargs) [ 502.174110] env[63531]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 502.174110] env[63531]: return func(*args, **kwargs) [ 502.174110] env[63531]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 502.174110] env[63531]: result = fn(*args, **kwargs) [ 502.174110] env[63531]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 502.174110] env[63531]: return f(*args, **kwargs) [ 502.174110] env[63531]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 502.174110] env[63531]: return db.service_get_minimum_version(context, binaries) [ 502.174110] env[63531]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 502.174110] env[63531]: _check_db_access() [ 502.174110] env[63531]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 502.174110] env[63531]: stacktrace = ''.join(traceback.format_stack()) [ 502.174110] env[63531]: [ 502.175078] env[63531]: ERROR nova.db.main.api [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 502.175078] env[63531]: result = function(*args, **kwargs) [ 502.175078] env[63531]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 502.175078] env[63531]: return func(*args, **kwargs) [ 502.175078] env[63531]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 502.175078] env[63531]: result = fn(*args, **kwargs) [ 502.175078] env[63531]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 502.175078] env[63531]: return f(*args, **kwargs) [ 502.175078] env[63531]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 502.175078] env[63531]: return db.service_get_minimum_version(context, binaries) [ 502.175078] env[63531]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 502.175078] env[63531]: _check_db_access() [ 502.175078] env[63531]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 502.175078] env[63531]: stacktrace = ''.join(traceback.format_stack()) [ 502.175078] env[63531]: [ 502.175490] env[63531]: WARNING nova.objects.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 502.175586] env[63531]: WARNING nova.objects.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] Failed to get minimum service version for cell 82819688-61a4-4a0c-802c-fcfc75bd49b7 [ 502.176039] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] Acquiring lock "singleton_lock" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 502.176204] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] Acquired lock "singleton_lock" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 502.176444] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] Releasing lock "singleton_lock" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 502.176760] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] Full set of CONF: {{(pid=63531) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 502.176904] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] ******************************************************************************** {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 502.177041] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] Configuration options gathered from: {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 502.177180] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 502.177373] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 502.177500] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] ================================================================================ {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 502.177706] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] allow_resize_to_same_host = True {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.177878] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] arq_binding_timeout = 300 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.178039] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] backdoor_port = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.178185] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] backdoor_socket = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.178353] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] block_device_allocate_retries = 60 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.178513] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] block_device_allocate_retries_interval = 3 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.178679] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cert = self.pem {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.178845] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.179014] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] compute_monitors = [] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.179191] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] config_dir = [] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.179358] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] config_drive_format = iso9660 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.179496] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.179665] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] config_source = [] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.179831] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] console_host = devstack {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.179993] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] control_exchange = nova {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.180169] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cpu_allocation_ratio = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.180329] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] daemon = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.180492] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] debug = True {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.180647] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] default_access_ip_network_name = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.180810] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] default_availability_zone = nova {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.180966] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] default_ephemeral_format = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.181133] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] default_green_pool_size = 1000 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.181365] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.181526] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] default_schedule_zone = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.181684] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] disk_allocation_ratio = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.181846] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] enable_new_services = True {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.182031] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] enabled_apis = ['osapi_compute'] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.182198] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] enabled_ssl_apis = [] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.182357] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] flat_injected = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.182513] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] force_config_drive = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.182667] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] force_raw_images = True {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.182834] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] graceful_shutdown_timeout = 5 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.182994] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] heal_instance_info_cache_interval = 60 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.183232] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] host = cpu-1 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.183388] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] initial_cpu_allocation_ratio = 4.0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.183546] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] initial_disk_allocation_ratio = 1.0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.183723] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] initial_ram_allocation_ratio = 1.0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.183946] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.184126] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] instance_build_timeout = 0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.184286] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] instance_delete_interval = 300 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.184452] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] instance_format = [instance: %(uuid)s] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.184614] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] instance_name_template = instance-%08x {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.184778] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] instance_usage_audit = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.184943] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] instance_usage_audit_period = month {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.185118] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.185280] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] instances_path = /opt/stack/data/nova/instances {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.185439] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] internal_service_availability_zone = internal {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.185595] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] key = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.185782] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] live_migration_retry_count = 30 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.186417] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] log_color = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.186417] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] log_config_append = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.186417] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.186417] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] log_dir = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.186570] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] log_file = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.186700] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] log_options = True {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.186857] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] log_rotate_interval = 1 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.187021] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] log_rotate_interval_type = days {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.187187] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] log_rotation_type = none {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.187316] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.187441] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.187603] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.187761] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.187889] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.188077] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] long_rpc_timeout = 1800 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.188248] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] max_concurrent_builds = 10 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.188404] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] max_concurrent_live_migrations = 1 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.188558] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] max_concurrent_snapshots = 5 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.188716] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] max_local_block_devices = 3 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.188875] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] max_logfile_count = 30 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.189037] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] max_logfile_size_mb = 200 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.189198] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] maximum_instance_delete_attempts = 5 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.189358] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] metadata_listen = 0.0.0.0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.189520] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] metadata_listen_port = 8775 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.189687] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] metadata_workers = 2 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.189847] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] migrate_max_retries = -1 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.190016] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] mkisofs_cmd = genisoimage {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.190225] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] my_block_storage_ip = 10.180.1.21 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.190354] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] my_ip = 10.180.1.21 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.190511] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] network_allocate_retries = 0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.190685] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.190849] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] osapi_compute_listen = 0.0.0.0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.191025] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] osapi_compute_listen_port = 8774 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.191185] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] osapi_compute_unique_server_name_scope = {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.191349] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] osapi_compute_workers = 2 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.191511] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] password_length = 12 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.191668] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] periodic_enable = True {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.191837] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] periodic_fuzzy_delay = 60 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.191991] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] pointer_model = usbtablet {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.192167] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] preallocate_images = none {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.192324] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] publish_errors = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.192454] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] pybasedir = /opt/stack/nova {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.192608] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] ram_allocation_ratio = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.192764] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] rate_limit_burst = 0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.192928] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] rate_limit_except_level = CRITICAL {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.193096] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] rate_limit_interval = 0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.193257] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] reboot_timeout = 0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.193415] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] reclaim_instance_interval = 0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.193565] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] record = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.193750] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] reimage_timeout_per_gb = 60 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.193929] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] report_interval = 120 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.194103] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] rescue_timeout = 0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.194267] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] reserved_host_cpus = 0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.194430] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] reserved_host_disk_mb = 0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.194590] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] reserved_host_memory_mb = 512 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.194749] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] reserved_huge_pages = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.194910] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] resize_confirm_window = 0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.195078] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] resize_fs_using_block_device = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.195238] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] resume_guests_state_on_host_boot = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.195402] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.195561] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] rpc_response_timeout = 60 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.195744] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] run_external_periodic_tasks = True {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.195925] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] running_deleted_instance_action = reap {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.196098] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] running_deleted_instance_poll_interval = 1800 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.196262] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] running_deleted_instance_timeout = 0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.196419] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] scheduler_instance_sync_interval = 120 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.196585] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] service_down_time = 720 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.196752] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] servicegroup_driver = db {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.196907] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] shell_completion = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.197076] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] shelved_offload_time = 0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.197237] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] shelved_poll_interval = 3600 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.197403] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] shutdown_timeout = 0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.197562] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] source_is_ipv6 = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.197719] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] ssl_only = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.197959] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.198161] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] sync_power_state_interval = 600 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.198326] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] sync_power_state_pool_size = 1000 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.198492] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] syslog_log_facility = LOG_USER {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.198651] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] tempdir = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.198810] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] timeout_nbd = 10 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.198976] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] transport_url = **** {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.199151] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] update_resources_interval = 0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.199311] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] use_cow_images = True {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.199468] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] use_eventlog = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.199624] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] use_journal = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.199781] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] use_json = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.199936] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] use_rootwrap_daemon = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.200121] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] use_stderr = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.200335] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] use_syslog = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.200498] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vcpu_pin_set = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.200667] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vif_plugging_is_fatal = True {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.200835] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vif_plugging_timeout = 300 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.200999] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] virt_mkfs = [] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.201177] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] volume_usage_poll_interval = 0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.201338] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] watch_log_file = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.201504] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] web = /usr/share/spice-html5 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.201684] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] os_brick.lock_path = /opt/stack/data/n-cpu-1 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.201851] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.202038] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] os_brick.wait_mpath_device_interval = 1 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.202196] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_concurrency.disable_process_locking = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.202738] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.202933] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.203119] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.203297] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_messaging_metrics.metrics_process_name = {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.203471] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.203638] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.203848] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] api.auth_strategy = keystone {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.204032] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] api.compute_link_prefix = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.204214] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.204388] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] api.dhcp_domain = novalocal {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.204557] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] api.enable_instance_password = True {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.204724] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] api.glance_link_prefix = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.204894] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.205078] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] api.instance_list_cells_batch_strategy = distributed {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.205247] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] api.instance_list_per_project_cells = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.205410] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] api.list_records_by_skipping_down_cells = True {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.205575] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] api.local_metadata_per_cell = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.205768] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] api.max_limit = 1000 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.205948] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] api.metadata_cache_expiration = 15 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.206138] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] api.neutron_default_tenant_id = default {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.206309] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] api.response_validation = warn {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.206627] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] api.use_neutron_default_nets = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.206687] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.206803] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] api.vendordata_dynamic_failure_fatal = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.206971] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.207158] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] api.vendordata_dynamic_ssl_certfile = {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.207330] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] api.vendordata_dynamic_targets = [] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.207494] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] api.vendordata_jsonfile_path = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.207677] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] api.vendordata_providers = ['StaticJSON'] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.207870] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cache.backend = dogpile.cache.memcached {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.208065] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cache.backend_argument = **** {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.208253] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cache.config_prefix = cache.oslo {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.208423] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cache.dead_timeout = 60.0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.208586] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cache.debug_cache_backend = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.208749] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cache.enable_retry_client = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.208911] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cache.enable_socket_keepalive = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.209092] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cache.enabled = True {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.209259] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cache.enforce_fips_mode = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.209421] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cache.expiration_time = 600 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.209585] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cache.hashclient_retry_attempts = 2 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.209751] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cache.hashclient_retry_delay = 1.0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.209916] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cache.memcache_dead_retry = 300 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.210085] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cache.memcache_password = **** {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.210249] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.210411] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.210571] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cache.memcache_pool_maxsize = 10 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.210734] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cache.memcache_pool_unused_timeout = 60 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.210896] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cache.memcache_sasl_enabled = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.211080] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cache.memcache_servers = ['localhost:11211'] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.211246] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cache.memcache_socket_timeout = 1.0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.211404] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cache.memcache_username = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.211570] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cache.proxies = [] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.211732] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cache.redis_db = 0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.211894] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cache.redis_password = **** {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.212130] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cache.redis_sentinel_service_name = mymaster {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.212388] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.212612] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cache.redis_server = localhost:6379 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.212801] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cache.redis_socket_timeout = 1.0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.212968] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cache.redis_username = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.213154] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cache.retry_attempts = 2 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.213319] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cache.retry_delay = 0.0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.213513] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cache.socket_keepalive_count = 1 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.213641] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cache.socket_keepalive_idle = 1 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.213831] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cache.socket_keepalive_interval = 1 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.213998] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cache.tls_allowed_ciphers = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.214192] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cache.tls_cafile = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.214366] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cache.tls_certfile = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.214727] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cache.tls_enabled = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.214953] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cache.tls_keyfile = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.215162] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cinder.auth_section = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.215344] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cinder.auth_type = password {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.215511] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cinder.cafile = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.215702] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cinder.catalog_info = volumev3::publicURL {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.215882] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cinder.certfile = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.216059] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cinder.collect_timing = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.216229] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cinder.cross_az_attach = True {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.216391] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cinder.debug = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.216549] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cinder.endpoint_template = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.216716] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cinder.http_retries = 3 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.216880] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cinder.insecure = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.217048] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cinder.keyfile = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.217222] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cinder.os_region_name = RegionOne {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.217388] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cinder.split_loggers = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.217549] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cinder.timeout = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.217725] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.217886] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] compute.cpu_dedicated_set = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.218097] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] compute.cpu_shared_set = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.218311] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] compute.image_type_exclude_list = [] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.218519] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] compute.live_migration_wait_for_vif_plug = True {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.218725] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] compute.max_concurrent_disk_ops = 0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.218939] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] compute.max_disk_devices_to_attach = -1 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.219134] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.219307] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.219471] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] compute.resource_provider_association_refresh = 300 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.219634] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.219797] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] compute.shutdown_retry_interval = 10 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.219975] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.220164] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] conductor.workers = 2 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.220342] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] console.allowed_origins = [] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.220501] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] console.ssl_ciphers = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.220669] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] console.ssl_minimum_version = default {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.220838] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] consoleauth.enforce_session_timeout = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.221011] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] consoleauth.token_ttl = 600 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.221186] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cyborg.cafile = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.221346] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cyborg.certfile = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.221507] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cyborg.collect_timing = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.221665] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cyborg.connect_retries = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.221825] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cyborg.connect_retry_delay = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.221983] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cyborg.endpoint_override = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.222156] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cyborg.insecure = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.222315] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cyborg.keyfile = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.222469] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cyborg.max_version = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.222624] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cyborg.min_version = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.222780] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cyborg.region_name = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.222940] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cyborg.retriable_status_codes = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.223109] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cyborg.service_name = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.223276] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cyborg.service_type = accelerator {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.223435] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cyborg.split_loggers = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.223613] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cyborg.status_code_retries = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.223757] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cyborg.status_code_retry_delay = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.223909] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cyborg.timeout = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.224100] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.224262] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] cyborg.version = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.224440] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] database.backend = sqlalchemy {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.224608] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] database.connection = **** {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.224774] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] database.connection_debug = 0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.224943] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] database.connection_parameters = {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.225117] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] database.connection_recycle_time = 3600 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.225280] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] database.connection_trace = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.225440] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] database.db_inc_retry_interval = True {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.225602] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] database.db_max_retries = 20 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.225792] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] database.db_max_retry_interval = 10 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.225968] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] database.db_retry_interval = 1 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.226142] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] database.max_overflow = 50 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.226304] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] database.max_pool_size = 5 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.226463] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] database.max_retries = 10 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.226628] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] database.mysql_sql_mode = TRADITIONAL {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.226801] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] database.mysql_wsrep_sync_wait = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.226972] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] database.pool_timeout = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.227152] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] database.retry_interval = 10 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.227310] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] database.slave_connection = **** {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.227468] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] database.sqlite_synchronous = True {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.227628] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] database.use_db_reconnect = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.227806] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] api_database.backend = sqlalchemy {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.227969] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] api_database.connection = **** {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.228184] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] api_database.connection_debug = 0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.228363] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] api_database.connection_parameters = {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.228532] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] api_database.connection_recycle_time = 3600 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.228697] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] api_database.connection_trace = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.228860] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] api_database.db_inc_retry_interval = True {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.229031] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] api_database.db_max_retries = 20 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.229206] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] api_database.db_max_retry_interval = 10 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.229364] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] api_database.db_retry_interval = 1 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.229526] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] api_database.max_overflow = 50 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.229686] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] api_database.max_pool_size = 5 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.229846] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] api_database.max_retries = 10 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.230024] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.230181] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] api_database.mysql_wsrep_sync_wait = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.230339] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] api_database.pool_timeout = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.230500] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] api_database.retry_interval = 10 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.230656] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] api_database.slave_connection = **** {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.230816] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] api_database.sqlite_synchronous = True {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.230993] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] devices.enabled_mdev_types = [] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.231183] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.231352] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] ephemeral_storage_encryption.default_format = luks {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.231513] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] ephemeral_storage_encryption.enabled = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.231675] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] ephemeral_storage_encryption.key_size = 512 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.231847] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] glance.api_servers = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.232014] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] glance.cafile = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.232180] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] glance.certfile = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.232346] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] glance.collect_timing = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.232503] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] glance.connect_retries = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.232658] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] glance.connect_retry_delay = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.232818] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] glance.debug = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.232983] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] glance.default_trusted_certificate_ids = [] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.233156] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] glance.enable_certificate_validation = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.233342] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] glance.enable_rbd_download = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.233518] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] glance.endpoint_override = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.233733] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] glance.insecure = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.233844] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] glance.keyfile = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.234060] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] glance.max_version = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.234243] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] glance.min_version = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.234410] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] glance.num_retries = 3 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.234577] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] glance.rbd_ceph_conf = {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.234739] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] glance.rbd_connect_timeout = 5 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.234907] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] glance.rbd_pool = {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.235082] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] glance.rbd_user = {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.235243] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] glance.region_name = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.235401] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] glance.retriable_status_codes = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.235557] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] glance.service_name = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.235748] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] glance.service_type = image {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.235923] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] glance.split_loggers = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.236101] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] glance.status_code_retries = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.236262] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] glance.status_code_retry_delay = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.236417] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] glance.timeout = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.236594] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.236799] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] glance.verify_glance_signatures = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.236975] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] glance.version = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.237155] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] guestfs.debug = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.237321] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] mks.enabled = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.237683] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.237874] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] image_cache.manager_interval = 2400 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.238050] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] image_cache.precache_concurrency = 1 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.238225] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] image_cache.remove_unused_base_images = True {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.238394] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.238559] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.238735] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] image_cache.subdirectory_name = _base {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.238916] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] ironic.api_max_retries = 60 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.239099] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] ironic.api_retry_interval = 2 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.239266] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] ironic.auth_section = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.239429] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] ironic.auth_type = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.239586] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] ironic.cafile = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.239743] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] ironic.certfile = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.239910] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] ironic.collect_timing = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.240079] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] ironic.conductor_group = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.240241] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] ironic.connect_retries = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.240399] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] ironic.connect_retry_delay = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.240557] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] ironic.endpoint_override = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.240720] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] ironic.insecure = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.240878] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] ironic.keyfile = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.241045] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] ironic.max_version = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.241208] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] ironic.min_version = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.241373] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] ironic.peer_list = [] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.241530] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] ironic.region_name = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.241688] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] ironic.retriable_status_codes = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.241852] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] ironic.serial_console_state_timeout = 10 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.242015] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] ironic.service_name = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.242189] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] ironic.service_type = baremetal {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.242349] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] ironic.shard = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.242514] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] ironic.split_loggers = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.242672] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] ironic.status_code_retries = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.242832] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] ironic.status_code_retry_delay = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.242991] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] ironic.timeout = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.243185] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.243347] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] ironic.version = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.243525] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.243699] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] key_manager.fixed_key = **** {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.243882] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.244049] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] barbican.barbican_api_version = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.244213] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] barbican.barbican_endpoint = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.244383] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] barbican.barbican_endpoint_type = public {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.244542] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] barbican.barbican_region_name = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.244699] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] barbican.cafile = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.244854] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] barbican.certfile = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.245017] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] barbican.collect_timing = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.245180] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] barbican.insecure = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.245337] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] barbican.keyfile = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.245497] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] barbican.number_of_retries = 60 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.245655] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] barbican.retry_delay = 1 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.245843] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] barbican.send_service_user_token = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.246020] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] barbican.split_loggers = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.246178] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] barbican.timeout = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.246337] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] barbican.verify_ssl = True {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.246491] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] barbican.verify_ssl_path = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.246657] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] barbican_service_user.auth_section = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.246872] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] barbican_service_user.auth_type = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.247064] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] barbican_service_user.cafile = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.247228] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] barbican_service_user.certfile = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.247390] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] barbican_service_user.collect_timing = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.247549] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] barbican_service_user.insecure = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.247704] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] barbican_service_user.keyfile = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.247865] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] barbican_service_user.split_loggers = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.248027] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] barbican_service_user.timeout = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.248196] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vault.approle_role_id = **** {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.248352] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vault.approle_secret_id = **** {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.248517] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vault.kv_mountpoint = secret {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.248672] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vault.kv_path = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.248833] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vault.kv_version = 2 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.248987] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vault.namespace = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.249152] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vault.root_token_id = **** {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.249305] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vault.ssl_ca_crt_file = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.249470] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vault.timeout = 60.0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.249629] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vault.use_ssl = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.249794] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.249972] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] keystone.auth_section = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.250147] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] keystone.auth_type = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.250305] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] keystone.cafile = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.250464] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] keystone.certfile = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.250625] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] keystone.collect_timing = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.250782] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] keystone.connect_retries = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.250939] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] keystone.connect_retry_delay = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.251103] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] keystone.endpoint_override = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.251265] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] keystone.insecure = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.251419] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] keystone.keyfile = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.251573] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] keystone.max_version = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.251726] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] keystone.min_version = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.251885] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] keystone.region_name = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.252047] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] keystone.retriable_status_codes = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.252204] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] keystone.service_name = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.252369] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] keystone.service_type = identity {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.252561] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] keystone.split_loggers = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.252680] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] keystone.status_code_retries = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.252835] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] keystone.status_code_retry_delay = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.252987] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] keystone.timeout = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.253177] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.253334] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] keystone.version = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.253527] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.connection_uri = {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.253685] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.cpu_mode = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.253851] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.cpu_model_extra_flags = [] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.254024] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.cpu_models = [] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.254195] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.cpu_power_governor_high = performance {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.254359] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.cpu_power_governor_low = powersave {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.254518] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.cpu_power_management = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.254687] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.254849] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.device_detach_attempts = 8 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.255015] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.device_detach_timeout = 20 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.255180] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.disk_cachemodes = [] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.255335] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.disk_prefix = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.255494] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.enabled_perf_events = [] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.255651] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.file_backed_memory = 0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.255851] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.gid_maps = [] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.256021] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.hw_disk_discard = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.256185] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.hw_machine_type = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.256353] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.images_rbd_ceph_conf = {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.256517] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.256691] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.256897] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.images_rbd_glance_store_name = {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.257092] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.images_rbd_pool = rbd {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.257265] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.images_type = default {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.257424] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.images_volume_group = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.257586] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.inject_key = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.257749] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.inject_partition = -2 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.257912] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.inject_password = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.258085] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.iscsi_iface = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.258251] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.iser_use_multipath = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.258411] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.live_migration_bandwidth = 0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.258572] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.live_migration_completion_timeout = 800 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.258731] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.live_migration_downtime = 500 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.258894] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.live_migration_downtime_delay = 75 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.259064] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.live_migration_downtime_steps = 10 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.259225] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.live_migration_inbound_addr = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.259386] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.live_migration_permit_auto_converge = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.259551] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.live_migration_permit_post_copy = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.259713] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.live_migration_scheme = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.259887] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.live_migration_timeout_action = abort {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.260059] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.live_migration_tunnelled = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.260221] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.live_migration_uri = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.260383] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.live_migration_with_native_tls = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.260543] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.max_queues = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.260705] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.mem_stats_period_seconds = 10 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.260953] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.261132] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.nfs_mount_options = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.261436] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.261609] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.num_aoe_discover_tries = 3 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.261776] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.num_iser_scan_tries = 5 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.261942] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.num_memory_encrypted_guests = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.262117] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.num_nvme_discover_tries = 5 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.262281] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.num_pcie_ports = 0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.262445] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.num_volume_scan_tries = 5 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.262614] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.pmem_namespaces = [] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.262769] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.quobyte_client_cfg = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.263076] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.263251] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.rbd_connect_timeout = 5 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.263417] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.263580] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.263741] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.rbd_secret_uuid = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.263907] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.rbd_user = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.264072] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.realtime_scheduler_priority = 1 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.264243] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.remote_filesystem_transport = ssh {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.264400] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.rescue_image_id = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.264559] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.rescue_kernel_id = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.264714] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.rescue_ramdisk_id = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.264884] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.rng_dev_path = /dev/urandom {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.265050] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.rx_queue_size = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.265220] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.smbfs_mount_options = {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.265491] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.265665] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.snapshot_compression = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.265847] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.snapshot_image_format = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.266076] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.266249] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.sparse_logical_volumes = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.266414] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.swtpm_enabled = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.266583] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.swtpm_group = tss {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.266777] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.swtpm_user = tss {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.266972] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.sysinfo_serial = unique {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.267147] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.tb_cache_size = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.267307] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.tx_queue_size = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.267474] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.uid_maps = [] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.267639] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.use_virtio_for_bridges = True {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.267810] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.virt_type = kvm {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.267982] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.volume_clear = zero {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.268159] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.volume_clear_size = 0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.268323] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.volume_use_multipath = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.268480] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.vzstorage_cache_path = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.268646] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.268812] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.vzstorage_mount_group = qemu {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.268981] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.vzstorage_mount_opts = [] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.269162] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.269434] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.269608] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.vzstorage_mount_user = stack {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.269774] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.269950] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] neutron.auth_section = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.270134] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] neutron.auth_type = password {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.270293] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] neutron.cafile = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.270452] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] neutron.certfile = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.270616] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] neutron.collect_timing = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.270777] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] neutron.connect_retries = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.270938] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] neutron.connect_retry_delay = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.271118] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] neutron.default_floating_pool = public {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.271278] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] neutron.endpoint_override = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.271440] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] neutron.extension_sync_interval = 600 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.271600] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] neutron.http_retries = 3 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.271762] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] neutron.insecure = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.271920] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] neutron.keyfile = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.272088] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] neutron.max_version = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.272260] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] neutron.metadata_proxy_shared_secret = **** {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.272420] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] neutron.min_version = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.272587] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] neutron.ovs_bridge = br-int {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.272754] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] neutron.physnets = [] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.272914] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] neutron.region_name = RegionOne {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.273086] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] neutron.retriable_status_codes = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.273260] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] neutron.service_metadata_proxy = True {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.273420] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] neutron.service_name = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.273586] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] neutron.service_type = network {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.273750] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] neutron.split_loggers = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.273908] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] neutron.status_code_retries = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.274077] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] neutron.status_code_retry_delay = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.274239] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] neutron.timeout = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.274417] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.274577] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] neutron.version = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.274746] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] notifications.bdms_in_notifications = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.274923] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] notifications.default_level = INFO {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.275103] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] notifications.notification_format = unversioned {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.275267] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] notifications.notify_on_state_change = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.275444] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.275617] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] pci.alias = [] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.275814] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] pci.device_spec = [] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.275984] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] pci.report_in_placement = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.276170] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] placement.auth_section = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.276410] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] placement.auth_type = password {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.276617] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] placement.auth_url = http://10.180.1.21/identity {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.276812] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] placement.cafile = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.276980] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] placement.certfile = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.277158] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] placement.collect_timing = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.277320] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] placement.connect_retries = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.277478] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] placement.connect_retry_delay = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.277636] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] placement.default_domain_id = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.277794] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] placement.default_domain_name = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.277952] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] placement.domain_id = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.278120] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] placement.domain_name = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.278282] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] placement.endpoint_override = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.278442] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] placement.insecure = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.278600] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] placement.keyfile = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.278756] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] placement.max_version = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.278914] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] placement.min_version = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.279091] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] placement.password = **** {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.279253] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] placement.project_domain_id = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.279418] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] placement.project_domain_name = Default {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.279586] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] placement.project_id = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.279757] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] placement.project_name = service {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.279927] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] placement.region_name = RegionOne {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.280099] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] placement.retriable_status_codes = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.280262] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] placement.service_name = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.280430] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] placement.service_type = placement {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.280593] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] placement.split_loggers = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.280751] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] placement.status_code_retries = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.280911] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] placement.status_code_retry_delay = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.281078] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] placement.system_scope = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.281239] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] placement.timeout = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.281395] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] placement.trust_id = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.281549] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] placement.user_domain_id = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.281714] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] placement.user_domain_name = Default {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.281873] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] placement.user_id = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.282052] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] placement.username = nova {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.282237] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.282398] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] placement.version = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.282573] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] quota.cores = 20 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.282740] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] quota.count_usage_from_placement = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.282913] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.283091] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] quota.injected_file_content_bytes = 10240 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.283256] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] quota.injected_file_path_length = 255 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.283419] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] quota.injected_files = 5 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.283581] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] quota.instances = 10 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.283745] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] quota.key_pairs = 100 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.283909] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] quota.metadata_items = 128 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.284082] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] quota.ram = 51200 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.284262] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] quota.recheck_quota = True {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.284413] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] quota.server_group_members = 10 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.284575] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] quota.server_groups = 10 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.284749] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.284913] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.285083] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] scheduler.image_metadata_prefilter = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.285246] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.285407] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] scheduler.max_attempts = 3 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.285567] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] scheduler.max_placement_results = 1000 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.285756] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.285929] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] scheduler.query_placement_for_image_type_support = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.286104] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.286276] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] scheduler.workers = 2 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.286447] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.286616] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.286817] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.286991] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.287169] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.287332] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.287496] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.287693] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.287869] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] filter_scheduler.host_subset_size = 1 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.288042] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.288207] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] filter_scheduler.image_properties_default_architecture = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.288370] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.288534] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] filter_scheduler.isolated_hosts = [] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.288699] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] filter_scheduler.isolated_images = [] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.288865] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] filter_scheduler.max_instances_per_host = 50 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.289033] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.289198] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.289359] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] filter_scheduler.pci_in_placement = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.289520] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.289678] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.289840] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.289997] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.290171] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.290334] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.290494] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] filter_scheduler.track_instance_changes = True {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.290669] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.290838] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] metrics.required = True {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.291008] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] metrics.weight_multiplier = 1.0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.291178] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] metrics.weight_of_unavailable = -10000.0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.291344] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] metrics.weight_setting = [] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.291660] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.291834] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] serial_console.enabled = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.292017] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] serial_console.port_range = 10000:20000 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.292191] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.292358] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.292528] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] serial_console.serialproxy_port = 6083 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.292694] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] service_user.auth_section = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.292867] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] service_user.auth_type = password {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.293067] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] service_user.cafile = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.293197] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] service_user.certfile = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.293361] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] service_user.collect_timing = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.293521] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] service_user.insecure = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.293680] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] service_user.keyfile = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.293849] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] service_user.send_service_user_token = True {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.294023] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] service_user.split_loggers = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.294192] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] service_user.timeout = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.294359] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] spice.agent_enabled = True {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.294520] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] spice.enabled = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.294829] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.295032] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] spice.html5proxy_host = 0.0.0.0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.295208] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] spice.html5proxy_port = 6082 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.295370] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] spice.image_compression = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.295529] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] spice.jpeg_compression = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.295702] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] spice.playback_compression = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.295885] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] spice.require_secure = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.296070] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] spice.server_listen = 127.0.0.1 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.296242] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.296401] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] spice.streaming_mode = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.296556] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] spice.zlib_compression = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.296763] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] upgrade_levels.baseapi = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.296977] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] upgrade_levels.compute = auto {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.297157] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] upgrade_levels.conductor = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.297319] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] upgrade_levels.scheduler = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.297484] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vendordata_dynamic_auth.auth_section = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.297646] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vendordata_dynamic_auth.auth_type = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.297801] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vendordata_dynamic_auth.cafile = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.297958] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vendordata_dynamic_auth.certfile = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.298132] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vendordata_dynamic_auth.collect_timing = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.298292] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vendordata_dynamic_auth.insecure = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.298445] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vendordata_dynamic_auth.keyfile = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.298602] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vendordata_dynamic_auth.split_loggers = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.298758] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vendordata_dynamic_auth.timeout = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.298931] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vmware.api_retry_count = 10 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.299103] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vmware.ca_file = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.299277] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vmware.cache_prefix = devstack-image-cache {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.299442] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vmware.cluster_name = testcl1 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.299605] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vmware.connection_pool_size = 10 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.299763] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vmware.console_delay_seconds = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.299933] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vmware.datastore_regex = ^datastore.* {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.300150] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.300325] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vmware.host_password = **** {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.300493] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vmware.host_port = 443 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.300661] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vmware.host_username = administrator@vsphere.local {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.300830] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vmware.insecure = True {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.300995] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vmware.integration_bridge = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.301170] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vmware.maximum_objects = 100 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.301330] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vmware.pbm_default_policy = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.301491] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vmware.pbm_enabled = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.301648] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vmware.pbm_wsdl_location = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.301815] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.301975] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vmware.serial_port_proxy_uri = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.302144] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vmware.serial_port_service_uri = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.302315] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vmware.task_poll_interval = 0.5 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.302486] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vmware.use_linked_clone = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.302656] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vmware.vnc_keymap = en-us {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.302845] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vmware.vnc_port = 5900 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.303019] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vmware.vnc_port_total = 10000 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.303216] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vnc.auth_schemes = ['none'] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.303390] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vnc.enabled = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.303684] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.303868] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.304051] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vnc.novncproxy_port = 6080 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.304234] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vnc.server_listen = 127.0.0.1 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.304471] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.304589] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vnc.vencrypt_ca_certs = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.304725] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vnc.vencrypt_client_cert = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.304887] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vnc.vencrypt_client_key = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.305080] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.305234] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] workarounds.disable_deep_image_inspection = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.305393] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] workarounds.disable_fallback_pcpu_query = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.305552] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] workarounds.disable_group_policy_check_upcall = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.305731] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.305908] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] workarounds.disable_rootwrap = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.306081] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] workarounds.enable_numa_live_migration = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.306243] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.306404] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.306563] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] workarounds.handle_virt_lifecycle_events = True {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.306744] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] workarounds.libvirt_disable_apic = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.306938] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] workarounds.never_download_image_if_on_rbd = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.307116] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.307282] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.307442] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.307601] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.307760] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.307918] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.308085] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.308245] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.308410] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.308589] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.308762] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] wsgi.client_socket_timeout = 900 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.308931] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] wsgi.default_pool_size = 1000 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.309108] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] wsgi.keep_alive = True {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.309276] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] wsgi.max_header_line = 16384 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.309439] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] wsgi.secure_proxy_ssl_header = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.309597] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] wsgi.ssl_ca_file = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.309756] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] wsgi.ssl_cert_file = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.309918] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] wsgi.ssl_key_file = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.310093] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] wsgi.tcp_keepidle = 600 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.310274] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.310442] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] zvm.ca_file = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.310602] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] zvm.cloud_connector_url = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.310889] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.311075] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] zvm.reachable_timeout = 300 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.311261] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_policy.enforce_new_defaults = True {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.311643] env[63531]: WARNING oslo_config.cfg [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 502.311823] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_policy.enforce_scope = True {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.311999] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_policy.policy_default_rule = default {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.312197] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.312370] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_policy.policy_file = policy.yaml {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.312545] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.312709] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.312870] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.313076] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.313269] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.313426] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.313598] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.313775] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] profiler.connection_string = messaging:// {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.313942] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] profiler.enabled = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.314121] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] profiler.es_doc_type = notification {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.314287] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] profiler.es_scroll_size = 10000 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.314452] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] profiler.es_scroll_time = 2m {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.314614] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] profiler.filter_error_trace = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.314778] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] profiler.hmac_keys = **** {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.314944] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] profiler.sentinel_service_name = mymaster {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.315120] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] profiler.socket_timeout = 0.1 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.315283] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] profiler.trace_requests = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.315440] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] profiler.trace_sqlalchemy = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.315620] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] profiler_jaeger.process_tags = {} {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.315806] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] profiler_jaeger.service_name_prefix = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.315975] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] profiler_otlp.service_name_prefix = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.316154] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] remote_debug.host = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.316313] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] remote_debug.port = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.316494] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.316661] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.316854] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.317039] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.317205] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.317366] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.317524] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.317685] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.317846] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.318019] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_messaging_rabbit.hostname = devstack {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.318184] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.318355] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.318520] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.318691] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.318898] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.319083] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.319250] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.319425] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.319588] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.319750] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.319915] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.320088] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.320251] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.320415] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.320573] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.320734] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.320895] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.321061] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.321230] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.321396] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_messaging_rabbit.ssl = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.321563] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.321729] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.321890] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.322064] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.322232] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_messaging_rabbit.ssl_version = {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.322391] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.322573] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.322736] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_messaging_notifications.retry = -1 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.322917] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.323100] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_messaging_notifications.transport_url = **** {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.323274] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_limit.auth_section = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.323437] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_limit.auth_type = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.323595] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_limit.cafile = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.323752] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_limit.certfile = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.323913] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_limit.collect_timing = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.324083] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_limit.connect_retries = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.324244] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_limit.connect_retry_delay = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.324400] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_limit.endpoint_id = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.324556] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_limit.endpoint_override = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.324717] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_limit.insecure = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.324872] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_limit.keyfile = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.325035] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_limit.max_version = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.325194] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_limit.min_version = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.325350] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_limit.region_name = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.325509] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_limit.retriable_status_codes = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.325672] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_limit.service_name = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.325852] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_limit.service_type = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.326027] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_limit.split_loggers = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.326191] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_limit.status_code_retries = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.326348] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_limit.status_code_retry_delay = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.326504] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_limit.timeout = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.326671] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_limit.valid_interfaces = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.326861] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_limit.version = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.327049] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_reports.file_event_handler = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.327221] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_reports.file_event_handler_interval = 1 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.327384] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] oslo_reports.log_dir = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.327555] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.327715] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vif_plug_linux_bridge_privileged.group = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.327876] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.328050] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.328219] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.328376] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vif_plug_linux_bridge_privileged.user = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.328545] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.328704] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vif_plug_ovs_privileged.group = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.328861] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vif_plug_ovs_privileged.helper_command = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.329033] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.329200] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.329356] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] vif_plug_ovs_privileged.user = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.329526] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] os_vif_linux_bridge.flat_interface = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.329706] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.329880] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.330059] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.330232] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.330396] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.330563] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.330725] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] os_vif_linux_bridge.vlan_interface = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.330991] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.331205] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] os_vif_ovs.isolate_vif = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.331380] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.331549] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.331720] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.331890] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] os_vif_ovs.ovsdb_interface = native {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.332068] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] os_vif_ovs.per_port_bridge = False {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.332244] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] privsep_osbrick.capabilities = [21] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.332402] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] privsep_osbrick.group = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.332559] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] privsep_osbrick.helper_command = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.332723] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.332887] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] privsep_osbrick.thread_pool_size = 8 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.333054] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] privsep_osbrick.user = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.333227] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.333382] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] nova_sys_admin.group = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.333537] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] nova_sys_admin.helper_command = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.333697] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.333856] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] nova_sys_admin.thread_pool_size = 8 {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.334017] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] nova_sys_admin.user = None {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.334147] env[63531]: DEBUG oslo_service.service [None req-6e98e42e-7d84-451a-b249-67bf999b192a None None] ******************************************************************************** {{(pid=63531) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 502.334624] env[63531]: INFO nova.service [-] Starting compute node (version 30.1.0) [ 502.837941] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-504fcc11-45ed-4c58-8cd6-a1b4c9313d8a None None] Getting list of instances from cluster (obj){ [ 502.837941] env[63531]: value = "domain-c8" [ 502.837941] env[63531]: _type = "ClusterComputeResource" [ 502.837941] env[63531]: } {{(pid=63531) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 502.839146] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55a50082-346e-4699-bd79-72704b03d0f8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 502.848122] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-504fcc11-45ed-4c58-8cd6-a1b4c9313d8a None None] Got total of 0 instances {{(pid=63531) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 502.848641] env[63531]: WARNING nova.virt.vmwareapi.driver [None req-504fcc11-45ed-4c58-8cd6-a1b4c9313d8a None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 502.849110] env[63531]: INFO nova.virt.node [None req-504fcc11-45ed-4c58-8cd6-a1b4c9313d8a None None] Generated node identity 9996a314-a62e-4282-8252-f387e7e3fabb [ 502.849347] env[63531]: INFO nova.virt.node [None req-504fcc11-45ed-4c58-8cd6-a1b4c9313d8a None None] Wrote node identity 9996a314-a62e-4282-8252-f387e7e3fabb to /opt/stack/data/n-cpu-1/compute_id [ 503.352152] env[63531]: WARNING nova.compute.manager [None req-504fcc11-45ed-4c58-8cd6-a1b4c9313d8a None None] Compute nodes ['9996a314-a62e-4282-8252-f387e7e3fabb'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 504.357364] env[63531]: INFO nova.compute.manager [None req-504fcc11-45ed-4c58-8cd6-a1b4c9313d8a None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 505.362339] env[63531]: WARNING nova.compute.manager [None req-504fcc11-45ed-4c58-8cd6-a1b4c9313d8a None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 505.362794] env[63531]: DEBUG oslo_concurrency.lockutils [None req-504fcc11-45ed-4c58-8cd6-a1b4c9313d8a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 505.362794] env[63531]: DEBUG oslo_concurrency.lockutils [None req-504fcc11-45ed-4c58-8cd6-a1b4c9313d8a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 505.362994] env[63531]: DEBUG oslo_concurrency.lockutils [None req-504fcc11-45ed-4c58-8cd6-a1b4c9313d8a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 505.363088] env[63531]: DEBUG nova.compute.resource_tracker [None req-504fcc11-45ed-4c58-8cd6-a1b4c9313d8a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63531) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 505.364032] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8293a8d2-005f-425c-b2e1-2fdd9fc37acd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 505.373091] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-472d94de-7edc-45c2-8f76-e398993a2d48 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 505.386659] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca6c20df-d8df-4a9b-b1c0-3e38b5b26c9d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 505.393091] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20294b58-9751-419b-b793-f7da28c7bbb3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 505.421817] env[63531]: DEBUG nova.compute.resource_tracker [None req-504fcc11-45ed-4c58-8cd6-a1b4c9313d8a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181589MB free_disk=170GB free_vcpus=48 pci_devices=None {{(pid=63531) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 505.421956] env[63531]: DEBUG oslo_concurrency.lockutils [None req-504fcc11-45ed-4c58-8cd6-a1b4c9313d8a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 505.422160] env[63531]: DEBUG oslo_concurrency.lockutils [None req-504fcc11-45ed-4c58-8cd6-a1b4c9313d8a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 505.924711] env[63531]: WARNING nova.compute.resource_tracker [None req-504fcc11-45ed-4c58-8cd6-a1b4c9313d8a None None] No compute node record for cpu-1:9996a314-a62e-4282-8252-f387e7e3fabb: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host 9996a314-a62e-4282-8252-f387e7e3fabb could not be found. [ 506.429040] env[63531]: INFO nova.compute.resource_tracker [None req-504fcc11-45ed-4c58-8cd6-a1b4c9313d8a None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: 9996a314-a62e-4282-8252-f387e7e3fabb [ 507.937758] env[63531]: DEBUG nova.compute.resource_tracker [None req-504fcc11-45ed-4c58-8cd6-a1b4c9313d8a None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=63531) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 507.938153] env[63531]: DEBUG nova.compute.resource_tracker [None req-504fcc11-45ed-4c58-8cd6-a1b4c9313d8a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=63531) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 508.109995] env[63531]: INFO nova.scheduler.client.report [None req-504fcc11-45ed-4c58-8cd6-a1b4c9313d8a None None] [req-fffdbdfa-b352-4a9f-9edf-55379f5c59d5] Created resource provider record via placement API for resource provider with UUID 9996a314-a62e-4282-8252-f387e7e3fabb and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 508.125909] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62a30f50-81f8-4da8-ac22-6970c4d067fc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 508.133461] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86a9839c-2d75-4886-91b8-632da0f05bf6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 508.163318] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2cd3dfa-3e6a-416f-bdaa-6db34c5c3998 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 508.170140] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6b8184a-2f16-4400-8ed7-8dabe17363ba {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 508.182681] env[63531]: DEBUG nova.compute.provider_tree [None req-504fcc11-45ed-4c58-8cd6-a1b4c9313d8a None None] Updating inventory in ProviderTree for provider 9996a314-a62e-4282-8252-f387e7e3fabb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 508.718776] env[63531]: DEBUG nova.scheduler.client.report [None req-504fcc11-45ed-4c58-8cd6-a1b4c9313d8a None None] Updated inventory for provider 9996a314-a62e-4282-8252-f387e7e3fabb with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 508.719035] env[63531]: DEBUG nova.compute.provider_tree [None req-504fcc11-45ed-4c58-8cd6-a1b4c9313d8a None None] Updating resource provider 9996a314-a62e-4282-8252-f387e7e3fabb generation from 0 to 1 during operation: update_inventory {{(pid=63531) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 508.719180] env[63531]: DEBUG nova.compute.provider_tree [None req-504fcc11-45ed-4c58-8cd6-a1b4c9313d8a None None] Updating inventory in ProviderTree for provider 9996a314-a62e-4282-8252-f387e7e3fabb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 508.765183] env[63531]: DEBUG nova.compute.provider_tree [None req-504fcc11-45ed-4c58-8cd6-a1b4c9313d8a None None] Updating resource provider 9996a314-a62e-4282-8252-f387e7e3fabb generation from 1 to 2 during operation: update_traits {{(pid=63531) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 509.269411] env[63531]: DEBUG nova.compute.resource_tracker [None req-504fcc11-45ed-4c58-8cd6-a1b4c9313d8a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63531) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 509.269763] env[63531]: DEBUG oslo_concurrency.lockutils [None req-504fcc11-45ed-4c58-8cd6-a1b4c9313d8a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.847s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 509.269814] env[63531]: DEBUG nova.service [None req-504fcc11-45ed-4c58-8cd6-a1b4c9313d8a None None] Creating RPC server for service compute {{(pid=63531) start /opt/stack/nova/nova/service.py:186}} [ 509.282612] env[63531]: DEBUG nova.service [None req-504fcc11-45ed-4c58-8cd6-a1b4c9313d8a None None] Join ServiceGroup membership for this service compute {{(pid=63531) start /opt/stack/nova/nova/service.py:203}} [ 509.282810] env[63531]: DEBUG nova.servicegroup.drivers.db [None req-504fcc11-45ed-4c58-8cd6-a1b4c9313d8a None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=63531) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 519.285284] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._sync_power_states {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 519.788756] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Getting list of instances from cluster (obj){ [ 519.788756] env[63531]: value = "domain-c8" [ 519.788756] env[63531]: _type = "ClusterComputeResource" [ 519.788756] env[63531]: } {{(pid=63531) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 519.789978] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f2b1bb2-0c29-46eb-93a0-6712673375b6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.798809] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Got total of 0 instances {{(pid=63531) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 519.799088] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 519.799398] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Getting list of instances from cluster (obj){ [ 519.799398] env[63531]: value = "domain-c8" [ 519.799398] env[63531]: _type = "ClusterComputeResource" [ 519.799398] env[63531]: } {{(pid=63531) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 519.800265] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c4ad394-763d-4d06-966a-27b6e387636d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.807639] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Got total of 0 instances {{(pid=63531) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 540.310872] env[63531]: DEBUG oslo_concurrency.lockutils [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] Acquiring lock "ca9c41cf-b5e6-400b-9b6c-34a065722969" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 540.311189] env[63531]: DEBUG oslo_concurrency.lockutils [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] Lock "ca9c41cf-b5e6-400b-9b6c-34a065722969" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 540.815881] env[63531]: DEBUG nova.compute.manager [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 541.352833] env[63531]: DEBUG oslo_concurrency.lockutils [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.353112] env[63531]: DEBUG oslo_concurrency.lockutils [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 541.358120] env[63531]: INFO nova.compute.claims [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 542.419511] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8199c80-e71a-4f91-99a7-599cdbd8d5ba {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.430555] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2acfcc41-629c-4ed3-b705-f0464ce86ca0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.470924] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-016ab16b-d151-44a2-9a4b-60b7ab284fde {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.484135] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3892537b-c9f7-46dd-8d09-e48be701ed10 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.496752] env[63531]: DEBUG nova.compute.provider_tree [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 542.975324] env[63531]: DEBUG oslo_concurrency.lockutils [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Acquiring lock "4410b013-868a-41e7-9dfb-584fb2d9d142" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.975425] env[63531]: DEBUG oslo_concurrency.lockutils [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Lock "4410b013-868a-41e7-9dfb-584fb2d9d142" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 542.991336] env[63531]: DEBUG oslo_concurrency.lockutils [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] Acquiring lock "baa18517-333c-4ebc-8de6-e2fc9576b9fd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.992729] env[63531]: DEBUG oslo_concurrency.lockutils [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] Lock "baa18517-333c-4ebc-8de6-e2fc9576b9fd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 543.000121] env[63531]: DEBUG nova.scheduler.client.report [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 543.262126] env[63531]: DEBUG oslo_concurrency.lockutils [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] Acquiring lock "c552e749-33ec-4093-be8c-ed3784bdc2b5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.262126] env[63531]: DEBUG oslo_concurrency.lockutils [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] Lock "c552e749-33ec-4093-be8c-ed3784bdc2b5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 543.479898] env[63531]: DEBUG nova.compute.manager [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 543.489081] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Acquiring lock "d1636eb1-b862-47c9-8729-7a9d475ace01" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.489506] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Lock "d1636eb1-b862-47c9-8729-7a9d475ace01" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 543.494793] env[63531]: DEBUG nova.compute.manager [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 543.510239] env[63531]: DEBUG oslo_concurrency.lockutils [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.157s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 543.510824] env[63531]: DEBUG nova.compute.manager [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 543.767052] env[63531]: DEBUG nova.compute.manager [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 543.993761] env[63531]: DEBUG nova.compute.manager [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 544.016943] env[63531]: DEBUG nova.compute.utils [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 544.018579] env[63531]: DEBUG nova.compute.manager [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 544.021806] env[63531]: DEBUG nova.network.neutron [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 544.025503] env[63531]: DEBUG oslo_concurrency.lockutils [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 544.025823] env[63531]: DEBUG oslo_concurrency.lockutils [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.002s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 544.027436] env[63531]: INFO nova.compute.claims [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 544.032865] env[63531]: DEBUG oslo_concurrency.lockutils [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 544.294025] env[63531]: DEBUG oslo_concurrency.lockutils [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 544.524715] env[63531]: DEBUG nova.compute.manager [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 544.532383] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 544.794634] env[63531]: DEBUG nova.policy [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '14bfc9c406c646a99b47e333a4f0eb2b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5a89eba35b18445094e11b0e23a1244f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 545.163203] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a90b6c6-d376-427b-b5c9-6dcc0f7cb789 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.171986] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85e1fdf9-e91b-427a-84dd-3adeff811a15 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.212203] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb15f384-8e3c-45d0-90c4-b4509cd5f7c3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.219907] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-693e4066-ea6d-45df-bab0-177f6d181562 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.233188] env[63531]: DEBUG nova.compute.provider_tree [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 545.542171] env[63531]: DEBUG nova.compute.manager [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 545.597790] env[63531]: DEBUG nova.virt.hardware [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 545.597790] env[63531]: DEBUG nova.virt.hardware [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 545.597790] env[63531]: DEBUG nova.virt.hardware [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 545.599139] env[63531]: DEBUG nova.virt.hardware [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 545.599447] env[63531]: DEBUG nova.virt.hardware [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 545.599709] env[63531]: DEBUG nova.virt.hardware [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 545.600045] env[63531]: DEBUG nova.virt.hardware [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 545.600389] env[63531]: DEBUG nova.virt.hardware [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 545.600852] env[63531]: DEBUG nova.virt.hardware [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 545.601140] env[63531]: DEBUG nova.virt.hardware [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 545.601409] env[63531]: DEBUG nova.virt.hardware [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 545.602386] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dc55e2d-ca51-491d-b68c-2b861396e64f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.615125] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0385bd44-b0f1-4b72-9731-4d5b75ed15fd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.634897] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc1b82a2-0a77-465c-ac18-280f758f6ff1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.736176] env[63531]: DEBUG nova.scheduler.client.report [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 546.046792] env[63531]: DEBUG nova.network.neutron [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] Successfully created port: 81228b1a-3a5e-426b-8ed3-d6273c990b2f {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 546.177111] env[63531]: DEBUG oslo_concurrency.lockutils [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Acquiring lock "6e963227-64f7-414d-88ec-cee5d5770bba" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 546.177418] env[63531]: DEBUG oslo_concurrency.lockutils [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Lock "6e963227-64f7-414d-88ec-cee5d5770bba" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 546.244971] env[63531]: DEBUG oslo_concurrency.lockutils [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.218s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 546.244971] env[63531]: DEBUG nova.compute.manager [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 546.248642] env[63531]: DEBUG oslo_concurrency.lockutils [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.215s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 546.250583] env[63531]: INFO nova.compute.claims [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 546.682209] env[63531]: DEBUG nova.compute.manager [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 546.749787] env[63531]: DEBUG nova.compute.utils [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 546.755423] env[63531]: DEBUG nova.compute.manager [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 546.755734] env[63531]: DEBUG nova.network.neutron [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 547.010312] env[63531]: DEBUG nova.policy [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'be263a91428e409784ae751574f1186d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3dda52b9a27c4d378c56f6876674e53a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 547.234297] env[63531]: DEBUG oslo_concurrency.lockutils [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 547.259173] env[63531]: DEBUG nova.compute.manager [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 547.398888] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c969f940-690f-439b-b3bd-aa79be09d7f3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.408477] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d42865f-afef-4730-b5e7-8788ffc2fdca {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.446110] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0c4e4d3-5588-4d75-a687-0577c02ec0c9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.457584] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44a6529d-c2e7-4364-8d77-e1d13160b1e7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.475311] env[63531]: DEBUG nova.compute.provider_tree [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 547.762900] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] Acquiring lock "35e9c404-bad4-4f8d-8c79-fb7250abb363" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 547.763178] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] Lock "35e9c404-bad4-4f8d-8c79-fb7250abb363" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 547.978455] env[63531]: DEBUG nova.scheduler.client.report [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 548.271927] env[63531]: DEBUG nova.compute.manager [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 548.279884] env[63531]: DEBUG nova.compute.manager [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 548.315934] env[63531]: DEBUG nova.virt.hardware [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 548.316231] env[63531]: DEBUG nova.virt.hardware [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 548.316352] env[63531]: DEBUG nova.virt.hardware [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 548.316535] env[63531]: DEBUG nova.virt.hardware [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 548.316677] env[63531]: DEBUG nova.virt.hardware [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 548.316823] env[63531]: DEBUG nova.virt.hardware [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 548.317097] env[63531]: DEBUG nova.virt.hardware [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 548.317277] env[63531]: DEBUG nova.virt.hardware [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 548.317446] env[63531]: DEBUG nova.virt.hardware [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 548.317605] env[63531]: DEBUG nova.virt.hardware [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 548.317774] env[63531]: DEBUG nova.virt.hardware [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 548.319378] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-209e4744-2699-4717-8918-b9fc0836c766 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.330503] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23fd3eb5-7460-421f-95ab-8c5228a406f3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.486437] env[63531]: DEBUG oslo_concurrency.lockutils [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.238s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 548.486571] env[63531]: DEBUG nova.compute.manager [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 548.489114] env[63531]: DEBUG oslo_concurrency.lockutils [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.196s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 548.494019] env[63531]: INFO nova.compute.claims [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 548.520579] env[63531]: DEBUG nova.network.neutron [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] Successfully created port: 554f8db0-c7eb-4bbd-b6e3-2aa10b61f54b {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 548.804946] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 548.819175] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] Acquiring lock "5287b3ed-12ec-47fb-93a2-94038cb2e7b1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 548.819429] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] Lock "5287b3ed-12ec-47fb-93a2-94038cb2e7b1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 548.996153] env[63531]: DEBUG nova.compute.utils [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 548.997583] env[63531]: DEBUG nova.compute.manager [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Not allocating networking since 'none' was specified. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 549.325035] env[63531]: DEBUG nova.compute.manager [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 549.499218] env[63531]: DEBUG nova.compute.manager [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 549.757206] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4714006-e0ac-46b2-bff0-dd0939be3ca8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.769145] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a609952-7f0e-465d-bcb8-eed383b58463 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.817678] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89772d92-a560-47a5-a1e0-01fc60d8534a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.827239] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2706ae8c-61a4-4a23-92e6-332b32363fbd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.850720] env[63531]: DEBUG nova.compute.provider_tree [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 549.868899] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.983218] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Acquiring lock "5b86b63a-f6c1-4983-a3cc-3cde7a10d17b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.984874] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Lock "5b86b63a-f6c1-4983-a3cc-3cde7a10d17b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 550.359016] env[63531]: DEBUG nova.scheduler.client.report [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 550.486421] env[63531]: DEBUG nova.compute.manager [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] [instance: 5b86b63a-f6c1-4983-a3cc-3cde7a10d17b] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 550.511776] env[63531]: DEBUG nova.compute.manager [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 550.561024] env[63531]: DEBUG nova.virt.hardware [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 550.561554] env[63531]: DEBUG nova.virt.hardware [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 550.561685] env[63531]: DEBUG nova.virt.hardware [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 550.561954] env[63531]: DEBUG nova.virt.hardware [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 550.562215] env[63531]: DEBUG nova.virt.hardware [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 550.562466] env[63531]: DEBUG nova.virt.hardware [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 550.562806] env[63531]: DEBUG nova.virt.hardware [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 550.563109] env[63531]: DEBUG nova.virt.hardware [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 550.563407] env[63531]: DEBUG nova.virt.hardware [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 550.563681] env[63531]: DEBUG nova.virt.hardware [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 550.564056] env[63531]: DEBUG nova.virt.hardware [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 550.566182] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eacac0d4-9456-4694-a3dd-856e4d488547 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.578498] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2ff38d9-d37d-4089-8226-378ff6a78642 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.605756] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Instance VIF info [] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 550.621310] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 550.622468] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9a6ad7e6-d393-4a38-afcb-5aebcd6a1fdb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.635343] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Created folder: OpenStack in parent group-v4. [ 550.635665] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Creating folder: Project (352ba60e02984342b2b65c751b2d40bc). Parent ref: group-v244585. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 550.636491] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5c518cee-a4f6-4540-8d5d-fc4a5c705b2f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.647081] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Created folder: Project (352ba60e02984342b2b65c751b2d40bc) in parent group-v244585. [ 550.647369] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Creating folder: Instances. Parent ref: group-v244586. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 550.650760] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-38471f37-e77a-4191-9bc2-4aed584ea0b3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.663649] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Created folder: Instances in parent group-v244586. [ 550.663649] env[63531]: DEBUG oslo.service.loopingcall [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 550.663649] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 550.663649] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e34a7896-4489-4a32-96dd-442f51cda1b3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.680132] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 550.680132] env[63531]: value = "task-1117461" [ 550.680132] env[63531]: _type = "Task" [ 550.680132] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 550.693677] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117461, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 550.781780] env[63531]: ERROR nova.compute.manager [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 81228b1a-3a5e-426b-8ed3-d6273c990b2f, please check neutron logs for more information. [ 550.781780] env[63531]: ERROR nova.compute.manager Traceback (most recent call last): [ 550.781780] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 550.781780] env[63531]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 550.781780] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 550.781780] env[63531]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 550.781780] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 550.781780] env[63531]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 550.781780] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 550.781780] env[63531]: ERROR nova.compute.manager self.force_reraise() [ 550.781780] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 550.781780] env[63531]: ERROR nova.compute.manager raise self.value [ 550.781780] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 550.781780] env[63531]: ERROR nova.compute.manager updated_port = self._update_port( [ 550.781780] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 550.781780] env[63531]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 550.782322] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 550.782322] env[63531]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 550.782322] env[63531]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 81228b1a-3a5e-426b-8ed3-d6273c990b2f, please check neutron logs for more information. [ 550.782322] env[63531]: ERROR nova.compute.manager [ 550.782322] env[63531]: Traceback (most recent call last): [ 550.782322] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 550.782322] env[63531]: listener.cb(fileno) [ 550.782322] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 550.782322] env[63531]: result = function(*args, **kwargs) [ 550.782322] env[63531]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 550.782322] env[63531]: return func(*args, **kwargs) [ 550.782322] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 550.782322] env[63531]: raise e [ 550.782322] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 550.782322] env[63531]: nwinfo = self.network_api.allocate_for_instance( [ 550.782322] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 550.782322] env[63531]: created_port_ids = self._update_ports_for_instance( [ 550.782322] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 550.782322] env[63531]: with excutils.save_and_reraise_exception(): [ 550.782322] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 550.782322] env[63531]: self.force_reraise() [ 550.782322] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 550.782322] env[63531]: raise self.value [ 550.782322] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 550.782322] env[63531]: updated_port = self._update_port( [ 550.782322] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 550.782322] env[63531]: _ensure_no_port_binding_failure(port) [ 550.782322] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 550.782322] env[63531]: raise exception.PortBindingFailed(port_id=port['id']) [ 550.783240] env[63531]: nova.exception.PortBindingFailed: Binding failed for port 81228b1a-3a5e-426b-8ed3-d6273c990b2f, please check neutron logs for more information. [ 550.783240] env[63531]: Removing descriptor: 14 [ 550.785192] env[63531]: ERROR nova.compute.manager [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 81228b1a-3a5e-426b-8ed3-d6273c990b2f, please check neutron logs for more information. [ 550.785192] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] Traceback (most recent call last): [ 550.785192] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 550.785192] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] yield resources [ 550.785192] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 550.785192] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] self.driver.spawn(context, instance, image_meta, [ 550.785192] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 550.785192] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] self._vmops.spawn(context, instance, image_meta, injected_files, [ 550.785192] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 550.785192] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] vm_ref = self.build_virtual_machine(instance, [ 550.785192] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 550.785608] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] vif_infos = vmwarevif.get_vif_info(self._session, [ 550.785608] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 550.785608] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] for vif in network_info: [ 550.785608] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 550.785608] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] return self._sync_wrapper(fn, *args, **kwargs) [ 550.785608] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 550.785608] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] self.wait() [ 550.785608] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 550.785608] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] self[:] = self._gt.wait() [ 550.785608] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 550.785608] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] return self._exit_event.wait() [ 550.785608] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 550.785608] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] result = hub.switch() [ 550.785992] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 550.785992] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] return self.greenlet.switch() [ 550.785992] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 550.785992] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] result = function(*args, **kwargs) [ 550.785992] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 550.785992] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] return func(*args, **kwargs) [ 550.785992] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 550.785992] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] raise e [ 550.785992] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 550.785992] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] nwinfo = self.network_api.allocate_for_instance( [ 550.785992] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 550.785992] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] created_port_ids = self._update_ports_for_instance( [ 550.785992] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 550.786412] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] with excutils.save_and_reraise_exception(): [ 550.786412] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 550.786412] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] self.force_reraise() [ 550.786412] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 550.786412] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] raise self.value [ 550.786412] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 550.786412] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] updated_port = self._update_port( [ 550.786412] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 550.786412] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] _ensure_no_port_binding_failure(port) [ 550.786412] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 550.786412] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] raise exception.PortBindingFailed(port_id=port['id']) [ 550.786412] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] nova.exception.PortBindingFailed: Binding failed for port 81228b1a-3a5e-426b-8ed3-d6273c990b2f, please check neutron logs for more information. [ 550.786412] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] [ 550.786772] env[63531]: INFO nova.compute.manager [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] Terminating instance [ 550.790451] env[63531]: DEBUG oslo_concurrency.lockutils [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] Acquiring lock "refresh_cache-ca9c41cf-b5e6-400b-9b6c-34a065722969" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 550.790788] env[63531]: DEBUG oslo_concurrency.lockutils [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] Acquired lock "refresh_cache-ca9c41cf-b5e6-400b-9b6c-34a065722969" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 550.790875] env[63531]: DEBUG nova.network.neutron [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 550.864130] env[63531]: DEBUG oslo_concurrency.lockutils [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.375s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 550.864652] env[63531]: DEBUG nova.compute.manager [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 550.867612] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.336s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 550.870287] env[63531]: INFO nova.compute.claims [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 551.014837] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 551.025362] env[63531]: DEBUG nova.compute.manager [req-fdd64f62-ff65-4f32-8fac-43c87256fffe req-4e859b0c-8ea6-4d2c-9df1-1b685402e98d service nova] [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] Received event network-changed-81228b1a-3a5e-426b-8ed3-d6273c990b2f {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 551.025697] env[63531]: DEBUG nova.compute.manager [req-fdd64f62-ff65-4f32-8fac-43c87256fffe req-4e859b0c-8ea6-4d2c-9df1-1b685402e98d service nova] [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] Refreshing instance network info cache due to event network-changed-81228b1a-3a5e-426b-8ed3-d6273c990b2f. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 551.025697] env[63531]: DEBUG oslo_concurrency.lockutils [req-fdd64f62-ff65-4f32-8fac-43c87256fffe req-4e859b0c-8ea6-4d2c-9df1-1b685402e98d service nova] Acquiring lock "refresh_cache-ca9c41cf-b5e6-400b-9b6c-34a065722969" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 551.194880] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117461, 'name': CreateVM_Task} progress is 25%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 551.370531] env[63531]: DEBUG nova.network.neutron [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 551.375082] env[63531]: DEBUG nova.compute.utils [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 551.379851] env[63531]: DEBUG nova.compute.manager [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 551.380042] env[63531]: DEBUG nova.network.neutron [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 551.584314] env[63531]: DEBUG nova.policy [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '10d88f6115314f80b94bbb46ea5a3094', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3f3996afe2ee46e8bff3123dbd4790d6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 551.695184] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117461, 'name': CreateVM_Task} progress is 25%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 551.696750] env[63531]: DEBUG nova.network.neutron [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 551.880767] env[63531]: DEBUG nova.compute.manager [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 552.061613] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f394837-94d7-4179-9d86-17bbabe6b380 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.070546] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da43be70-f845-4e09-b41b-4ccb16e20994 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.110051] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41a02e33-7247-4886-83c9-9938120ca7de {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.118189] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-074c4306-3e0a-48ce-bf97-c1d651f53c5a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.142372] env[63531]: DEBUG nova.compute.provider_tree [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 552.194646] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117461, 'name': CreateVM_Task} progress is 25%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 552.199686] env[63531]: DEBUG oslo_concurrency.lockutils [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] Releasing lock "refresh_cache-ca9c41cf-b5e6-400b-9b6c-34a065722969" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 552.199835] env[63531]: DEBUG nova.compute.manager [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 552.200073] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 552.200362] env[63531]: DEBUG oslo_concurrency.lockutils [req-fdd64f62-ff65-4f32-8fac-43c87256fffe req-4e859b0c-8ea6-4d2c-9df1-1b685402e98d service nova] Acquired lock "refresh_cache-ca9c41cf-b5e6-400b-9b6c-34a065722969" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 552.200532] env[63531]: DEBUG nova.network.neutron [req-fdd64f62-ff65-4f32-8fac-43c87256fffe req-4e859b0c-8ea6-4d2c-9df1-1b685402e98d service nova] [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] Refreshing network info cache for port 81228b1a-3a5e-426b-8ed3-d6273c990b2f {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 552.201654] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a5e3a392-f4cf-4b00-aa41-d67bbf1fb125 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.213649] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f789e14c-fea0-4e12-8842-daf060ab1ea5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.238121] env[63531]: WARNING nova.virt.vmwareapi.vmops [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ca9c41cf-b5e6-400b-9b6c-34a065722969 could not be found. [ 552.238419] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 552.238724] env[63531]: INFO nova.compute.manager [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] Took 0.04 seconds to destroy the instance on the hypervisor. [ 552.238966] env[63531]: DEBUG oslo.service.loopingcall [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 552.240522] env[63531]: DEBUG nova.compute.manager [-] [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 552.240597] env[63531]: DEBUG nova.network.neutron [-] [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 552.287805] env[63531]: DEBUG nova.network.neutron [-] [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 552.305421] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Acquiring lock "6a380967-bca6-47d2-b052-988e6c1004ee" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.305421] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Lock "6a380967-bca6-47d2-b052-988e6c1004ee" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 552.647166] env[63531]: DEBUG nova.scheduler.client.report [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 552.700850] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117461, 'name': CreateVM_Task} progress is 25%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 552.711537] env[63531]: DEBUG nova.network.neutron [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] Successfully created port: b832b20c-e3b8-40a9-a4cf-91f8c961fd94 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 552.745245] env[63531]: DEBUG nova.network.neutron [req-fdd64f62-ff65-4f32-8fac-43c87256fffe req-4e859b0c-8ea6-4d2c-9df1-1b685402e98d service nova] [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 552.792756] env[63531]: DEBUG nova.network.neutron [-] [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 552.808576] env[63531]: DEBUG nova.compute.manager [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 552.894929] env[63531]: DEBUG nova.compute.manager [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 552.940876] env[63531]: DEBUG nova.virt.hardware [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 552.940876] env[63531]: DEBUG nova.virt.hardware [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 552.940876] env[63531]: DEBUG nova.virt.hardware [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 552.941362] env[63531]: DEBUG nova.virt.hardware [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 552.941362] env[63531]: DEBUG nova.virt.hardware [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 552.941362] env[63531]: DEBUG nova.virt.hardware [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 552.941362] env[63531]: DEBUG nova.virt.hardware [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 552.941527] env[63531]: DEBUG nova.virt.hardware [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 552.941657] env[63531]: DEBUG nova.virt.hardware [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 552.941811] env[63531]: DEBUG nova.virt.hardware [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 552.941977] env[63531]: DEBUG nova.virt.hardware [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 552.943510] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea5cced5-4916-4142-a6c7-7dc5e02102d4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.954809] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbd57953-0caf-40cf-bb82-dfbea7f05c46 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.152596] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.285s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 553.153189] env[63531]: DEBUG nova.compute.manager [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 553.155670] env[63531]: DEBUG oslo_concurrency.lockutils [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.922s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 553.157088] env[63531]: INFO nova.compute.claims [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 553.203933] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117461, 'name': CreateVM_Task} progress is 25%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 553.299599] env[63531]: INFO nova.compute.manager [-] [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] Took 1.06 seconds to deallocate network for instance. [ 553.302654] env[63531]: DEBUG nova.compute.claims [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] Aborting claim: {{(pid=63531) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 553.302959] env[63531]: DEBUG oslo_concurrency.lockutils [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 553.343015] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 553.348483] env[63531]: DEBUG nova.network.neutron [req-fdd64f62-ff65-4f32-8fac-43c87256fffe req-4e859b0c-8ea6-4d2c-9df1-1b685402e98d service nova] [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 553.471112] env[63531]: ERROR nova.compute.manager [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 554f8db0-c7eb-4bbd-b6e3-2aa10b61f54b, please check neutron logs for more information. [ 553.471112] env[63531]: ERROR nova.compute.manager Traceback (most recent call last): [ 553.471112] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 553.471112] env[63531]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 553.471112] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 553.471112] env[63531]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 553.471112] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 553.471112] env[63531]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 553.471112] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 553.471112] env[63531]: ERROR nova.compute.manager self.force_reraise() [ 553.471112] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 553.471112] env[63531]: ERROR nova.compute.manager raise self.value [ 553.471112] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 553.471112] env[63531]: ERROR nova.compute.manager updated_port = self._update_port( [ 553.471112] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 553.471112] env[63531]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 553.471680] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 553.471680] env[63531]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 553.471680] env[63531]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 554f8db0-c7eb-4bbd-b6e3-2aa10b61f54b, please check neutron logs for more information. [ 553.471680] env[63531]: ERROR nova.compute.manager [ 553.471680] env[63531]: Traceback (most recent call last): [ 553.471680] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 553.471680] env[63531]: listener.cb(fileno) [ 553.471680] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 553.471680] env[63531]: result = function(*args, **kwargs) [ 553.471680] env[63531]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 553.471680] env[63531]: return func(*args, **kwargs) [ 553.471680] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 553.471680] env[63531]: raise e [ 553.471680] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 553.471680] env[63531]: nwinfo = self.network_api.allocate_for_instance( [ 553.471680] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 553.471680] env[63531]: created_port_ids = self._update_ports_for_instance( [ 553.471680] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 553.471680] env[63531]: with excutils.save_and_reraise_exception(): [ 553.471680] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 553.471680] env[63531]: self.force_reraise() [ 553.471680] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 553.471680] env[63531]: raise self.value [ 553.471680] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 553.471680] env[63531]: updated_port = self._update_port( [ 553.471680] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 553.471680] env[63531]: _ensure_no_port_binding_failure(port) [ 553.471680] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 553.471680] env[63531]: raise exception.PortBindingFailed(port_id=port['id']) [ 553.472511] env[63531]: nova.exception.PortBindingFailed: Binding failed for port 554f8db0-c7eb-4bbd-b6e3-2aa10b61f54b, please check neutron logs for more information. [ 553.472511] env[63531]: Removing descriptor: 15 [ 553.472511] env[63531]: ERROR nova.compute.manager [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 554f8db0-c7eb-4bbd-b6e3-2aa10b61f54b, please check neutron logs for more information. [ 553.472511] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] Traceback (most recent call last): [ 553.472511] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 553.472511] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] yield resources [ 553.472511] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 553.472511] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] self.driver.spawn(context, instance, image_meta, [ 553.472511] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 553.472511] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 553.472511] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 553.472511] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] vm_ref = self.build_virtual_machine(instance, [ 553.472899] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 553.472899] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] vif_infos = vmwarevif.get_vif_info(self._session, [ 553.472899] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 553.472899] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] for vif in network_info: [ 553.472899] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 553.472899] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] return self._sync_wrapper(fn, *args, **kwargs) [ 553.472899] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 553.472899] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] self.wait() [ 553.472899] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 553.472899] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] self[:] = self._gt.wait() [ 553.472899] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 553.472899] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] return self._exit_event.wait() [ 553.472899] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 553.473301] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] result = hub.switch() [ 553.473301] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 553.473301] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] return self.greenlet.switch() [ 553.473301] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 553.473301] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] result = function(*args, **kwargs) [ 553.473301] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 553.473301] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] return func(*args, **kwargs) [ 553.473301] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 553.473301] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] raise e [ 553.473301] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 553.473301] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] nwinfo = self.network_api.allocate_for_instance( [ 553.473301] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 553.473301] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] created_port_ids = self._update_ports_for_instance( [ 553.473701] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 553.473701] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] with excutils.save_and_reraise_exception(): [ 553.473701] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 553.473701] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] self.force_reraise() [ 553.473701] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 553.473701] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] raise self.value [ 553.473701] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 553.473701] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] updated_port = self._update_port( [ 553.473701] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 553.473701] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] _ensure_no_port_binding_failure(port) [ 553.473701] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 553.473701] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] raise exception.PortBindingFailed(port_id=port['id']) [ 553.474140] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] nova.exception.PortBindingFailed: Binding failed for port 554f8db0-c7eb-4bbd-b6e3-2aa10b61f54b, please check neutron logs for more information. [ 553.474140] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] [ 553.474140] env[63531]: INFO nova.compute.manager [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] Terminating instance [ 553.476442] env[63531]: DEBUG oslo_concurrency.lockutils [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] Acquiring lock "refresh_cache-baa18517-333c-4ebc-8de6-e2fc9576b9fd" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 553.476442] env[63531]: DEBUG oslo_concurrency.lockutils [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] Acquired lock "refresh_cache-baa18517-333c-4ebc-8de6-e2fc9576b9fd" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 553.477143] env[63531]: DEBUG nova.network.neutron [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 553.663205] env[63531]: DEBUG nova.compute.utils [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 553.670105] env[63531]: DEBUG nova.compute.manager [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 553.670105] env[63531]: DEBUG nova.network.neutron [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 553.699687] env[63531]: DEBUG nova.compute.manager [req-f7976ec4-f393-4761-96da-cb6a50a86927 req-83f3ef59-2315-4efe-8a8d-f617aca8b3ac service nova] [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] Received event network-vif-deleted-81228b1a-3a5e-426b-8ed3-d6273c990b2f {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 553.699687] env[63531]: DEBUG nova.compute.manager [req-f7976ec4-f393-4761-96da-cb6a50a86927 req-83f3ef59-2315-4efe-8a8d-f617aca8b3ac service nova] [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] Received event network-changed-554f8db0-c7eb-4bbd-b6e3-2aa10b61f54b {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 553.699869] env[63531]: DEBUG nova.compute.manager [req-f7976ec4-f393-4761-96da-cb6a50a86927 req-83f3ef59-2315-4efe-8a8d-f617aca8b3ac service nova] [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] Refreshing instance network info cache due to event network-changed-554f8db0-c7eb-4bbd-b6e3-2aa10b61f54b. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 553.700056] env[63531]: DEBUG oslo_concurrency.lockutils [req-f7976ec4-f393-4761-96da-cb6a50a86927 req-83f3ef59-2315-4efe-8a8d-f617aca8b3ac service nova] Acquiring lock "refresh_cache-baa18517-333c-4ebc-8de6-e2fc9576b9fd" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 553.708131] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117461, 'name': CreateVM_Task} progress is 25%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 553.851748] env[63531]: DEBUG oslo_concurrency.lockutils [req-fdd64f62-ff65-4f32-8fac-43c87256fffe req-4e859b0c-8ea6-4d2c-9df1-1b685402e98d service nova] Releasing lock "refresh_cache-ca9c41cf-b5e6-400b-9b6c-34a065722969" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 553.949952] env[63531]: DEBUG nova.policy [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '979d9701753b4d0082abe8b5ef217529', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4cb35716050e4e8c95948ca7913bf815', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 554.070688] env[63531]: DEBUG nova.network.neutron [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 554.176321] env[63531]: DEBUG nova.compute.manager [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 554.211027] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117461, 'name': CreateVM_Task} progress is 25%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 554.375615] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3b39609-859d-4ae6-9e58-a16ffe034cc1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.384340] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f02b3793-bbe5-4552-8f7f-57d4df8dd3be {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.421262] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c2a846f-bb0d-4da5-aa18-23443bd2d2bf {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.429554] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaaa3f35-91c7-420b-974a-2277f7b83222 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.445475] env[63531]: DEBUG nova.compute.provider_tree [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 554.624306] env[63531]: DEBUG nova.network.neutron [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 554.710391] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117461, 'name': CreateVM_Task, 'duration_secs': 3.74623} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 554.710575] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 554.711784] env[63531]: DEBUG oslo_vmware.service [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd8b3def-d0d4-43a2-a3c3-05014ac90c49 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.719413] env[63531]: DEBUG oslo_concurrency.lockutils [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 554.719873] env[63531]: DEBUG oslo_concurrency.lockutils [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 554.720558] env[63531]: DEBUG oslo_concurrency.lockutils [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 554.720960] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dc4a36b0-af03-4a47-a43e-3f4069e53bcf {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.726826] env[63531]: DEBUG oslo_vmware.api [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Waiting for the task: (returnval){ [ 554.726826] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5226edae-f209-026d-4456-7cd0d0833694" [ 554.726826] env[63531]: _type = "Task" [ 554.726826] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 554.736543] env[63531]: DEBUG oslo_vmware.api [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5226edae-f209-026d-4456-7cd0d0833694, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 554.948879] env[63531]: DEBUG nova.scheduler.client.report [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 555.128290] env[63531]: DEBUG oslo_concurrency.lockutils [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] Releasing lock "refresh_cache-baa18517-333c-4ebc-8de6-e2fc9576b9fd" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 555.128290] env[63531]: DEBUG nova.compute.manager [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 555.128290] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 555.128290] env[63531]: DEBUG oslo_concurrency.lockutils [req-f7976ec4-f393-4761-96da-cb6a50a86927 req-83f3ef59-2315-4efe-8a8d-f617aca8b3ac service nova] Acquired lock "refresh_cache-baa18517-333c-4ebc-8de6-e2fc9576b9fd" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 555.128290] env[63531]: DEBUG nova.network.neutron [req-f7976ec4-f393-4761-96da-cb6a50a86927 req-83f3ef59-2315-4efe-8a8d-f617aca8b3ac service nova] [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] Refreshing network info cache for port 554f8db0-c7eb-4bbd-b6e3-2aa10b61f54b {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 555.133384] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-34c10f92-c222-451e-8c53-4da32f34acd9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.148218] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a3b497c-8422-4bcd-839d-05715c231adf {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.174726] env[63531]: WARNING nova.virt.vmwareapi.vmops [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance baa18517-333c-4ebc-8de6-e2fc9576b9fd could not be found. [ 555.174726] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 555.174914] env[63531]: INFO nova.compute.manager [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] Took 0.05 seconds to destroy the instance on the hypervisor. [ 555.175543] env[63531]: DEBUG oslo.service.loopingcall [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 555.175543] env[63531]: DEBUG nova.compute.manager [-] [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 555.175543] env[63531]: DEBUG nova.network.neutron [-] [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 555.187502] env[63531]: DEBUG nova.compute.manager [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 555.226012] env[63531]: DEBUG nova.network.neutron [-] [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 555.234171] env[63531]: DEBUG nova.virt.hardware [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 555.234391] env[63531]: DEBUG nova.virt.hardware [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 555.234548] env[63531]: DEBUG nova.virt.hardware [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 555.234736] env[63531]: DEBUG nova.virt.hardware [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 555.234928] env[63531]: DEBUG nova.virt.hardware [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 555.235028] env[63531]: DEBUG nova.virt.hardware [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 555.235239] env[63531]: DEBUG nova.virt.hardware [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 555.235394] env[63531]: DEBUG nova.virt.hardware [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 555.235553] env[63531]: DEBUG nova.virt.hardware [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 555.235951] env[63531]: DEBUG nova.virt.hardware [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 555.235951] env[63531]: DEBUG nova.virt.hardware [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 555.241541] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92b5b931-79fd-46d2-9fe4-e0078fa2b926 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.253456] env[63531]: DEBUG oslo_concurrency.lockutils [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 555.253658] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 555.253871] env[63531]: DEBUG oslo_concurrency.lockutils [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 555.254046] env[63531]: DEBUG oslo_concurrency.lockutils [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 555.254467] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 555.255112] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c66e6fce-c21f-4e8d-911d-8c1af9298240 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.260974] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03690935-96b3-4c07-a965-56ed8adbc77c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.279437] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 555.279574] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 555.280514] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c1db259-b4c5-46dc-bbf7-b9088253cb34 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.289949] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-084b399b-206b-4ef5-a706-404cff4f43c4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.297115] env[63531]: DEBUG oslo_vmware.api [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Waiting for the task: (returnval){ [ 555.297115] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5202a375-24ba-c15b-2893-00087009b349" [ 555.297115] env[63531]: _type = "Task" [ 555.297115] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 555.303234] env[63531]: DEBUG oslo_vmware.api [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5202a375-24ba-c15b-2893-00087009b349, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 555.454999] env[63531]: DEBUG oslo_concurrency.lockutils [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.299s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 555.455543] env[63531]: DEBUG nova.compute.manager [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 555.458324] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.654s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 555.460550] env[63531]: INFO nova.compute.claims [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 555.728594] env[63531]: DEBUG nova.network.neutron [req-f7976ec4-f393-4761-96da-cb6a50a86927 req-83f3ef59-2315-4efe-8a8d-f617aca8b3ac service nova] [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 555.732710] env[63531]: DEBUG nova.network.neutron [-] [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 555.806180] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Preparing fetch location {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 555.806454] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Creating directory with path [datastore1] vmware_temp/f265b2d4-8bb5-4203-8958-00950d4e1576/9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 555.806694] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-18414a75-1770-453e-8a15-dfde54954673 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.833320] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Created directory with path [datastore1] vmware_temp/f265b2d4-8bb5-4203-8958-00950d4e1576/9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 555.833929] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Fetch image to [datastore1] vmware_temp/f265b2d4-8bb5-4203-8958-00950d4e1576/9d550399-588c-4bd5-8436-eba9b523bd13/tmp-sparse.vmdk {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 555.834130] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Downloading image file data 9d550399-588c-4bd5-8436-eba9b523bd13 to [datastore1] vmware_temp/f265b2d4-8bb5-4203-8958-00950d4e1576/9d550399-588c-4bd5-8436-eba9b523bd13/tmp-sparse.vmdk on the data store datastore1 {{(pid=63531) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 555.839569] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e571b49-d72a-4c12-a9cb-a335ee738f02 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.853731] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d07a60b7-fcbb-4b5c-9055-422b04eb68b4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.868114] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81eac85b-722f-45e1-b359-50ea4116b9ab {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.905920] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb420e66-55e2-421c-b9f0-2047fdc2ce93 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.913445] env[63531]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-e5ba7283-f18c-4661-ad45-54d8e4ab0af7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.942280] env[63531]: DEBUG nova.virt.vmwareapi.images [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Downloading image file data 9d550399-588c-4bd5-8436-eba9b523bd13 to the data store datastore1 {{(pid=63531) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 555.969114] env[63531]: DEBUG nova.compute.utils [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 555.971257] env[63531]: DEBUG nova.compute.manager [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 555.971426] env[63531]: DEBUG nova.network.neutron [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 556.020038] env[63531]: DEBUG oslo_concurrency.lockutils [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] Acquiring lock "2ae517ae-a159-4dc6-8100-457a4cea1aeb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 556.020360] env[63531]: DEBUG oslo_concurrency.lockutils [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] Lock "2ae517ae-a159-4dc6-8100-457a4cea1aeb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 556.048041] env[63531]: DEBUG nova.policy [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '30c2b29064b44c48ad7e76d445c65317', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '04a4c814245346a9b73253a25a822fef', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 556.152036] env[63531]: DEBUG oslo_vmware.rw_handles [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/f265b2d4-8bb5-4203-8958-00950d4e1576/9d550399-588c-4bd5-8436-eba9b523bd13/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=63531) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 556.222257] env[63531]: DEBUG nova.network.neutron [req-f7976ec4-f393-4761-96da-cb6a50a86927 req-83f3ef59-2315-4efe-8a8d-f617aca8b3ac service nova] [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 556.239832] env[63531]: INFO nova.compute.manager [-] [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] Took 1.06 seconds to deallocate network for instance. [ 556.243018] env[63531]: DEBUG nova.compute.claims [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] Aborting claim: {{(pid=63531) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 556.243228] env[63531]: DEBUG oslo_concurrency.lockutils [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 556.472251] env[63531]: DEBUG nova.compute.manager [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 556.539026] env[63531]: DEBUG nova.network.neutron [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] Successfully created port: 73d60e7d-c4d2-475d-aeed-b12ea040e3df {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 556.726947] env[63531]: DEBUG oslo_concurrency.lockutils [req-f7976ec4-f393-4761-96da-cb6a50a86927 req-83f3ef59-2315-4efe-8a8d-f617aca8b3ac service nova] Releasing lock "refresh_cache-baa18517-333c-4ebc-8de6-e2fc9576b9fd" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 556.728122] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c4e9352-b41a-473f-a85c-e020b292a79f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.739476] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6ffa3c3-0e2c-4db3-b51d-997247efa2e2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.798500] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7cb0211-4b56-443c-a68f-48bbf2206ea2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.808252] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fb4dca4-a0a5-4dd3-8e6e-4ecd65e02398 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.825528] env[63531]: DEBUG nova.compute.provider_tree [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 556.852309] env[63531]: DEBUG oslo_vmware.rw_handles [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Completed reading data from the image iterator. {{(pid=63531) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 556.852577] env[63531]: DEBUG oslo_vmware.rw_handles [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Closing write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/f265b2d4-8bb5-4203-8958-00950d4e1576/9d550399-588c-4bd5-8436-eba9b523bd13/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=63531) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 557.004602] env[63531]: DEBUG nova.virt.vmwareapi.images [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Downloaded image file data 9d550399-588c-4bd5-8436-eba9b523bd13 to vmware_temp/f265b2d4-8bb5-4203-8958-00950d4e1576/9d550399-588c-4bd5-8436-eba9b523bd13/tmp-sparse.vmdk on the data store datastore1 {{(pid=63531) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 557.006338] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Caching image {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 557.006589] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Copying Virtual Disk [datastore1] vmware_temp/f265b2d4-8bb5-4203-8958-00950d4e1576/9d550399-588c-4bd5-8436-eba9b523bd13/tmp-sparse.vmdk to [datastore1] vmware_temp/f265b2d4-8bb5-4203-8958-00950d4e1576/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 557.006872] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f16efb6f-b463-4dd4-b462-904c7b6d0d11 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.016208] env[63531]: DEBUG oslo_vmware.api [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Waiting for the task: (returnval){ [ 557.016208] env[63531]: value = "task-1117462" [ 557.016208] env[63531]: _type = "Task" [ 557.016208] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 557.024228] env[63531]: DEBUG oslo_vmware.api [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': task-1117462, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 557.307387] env[63531]: DEBUG nova.compute.manager [req-0eba0727-74d0-4127-9f95-9ffd6e945ebb req-14ff517f-16eb-4207-8e75-e3b1211695d3 service nova] [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] Received event network-vif-deleted-554f8db0-c7eb-4bbd-b6e3-2aa10b61f54b {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 557.333228] env[63531]: DEBUG nova.scheduler.client.report [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 557.375173] env[63531]: DEBUG nova.network.neutron [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] Successfully created port: e70dcd41-edfa-4975-a540-6e7d7bbd6280 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 557.487755] env[63531]: DEBUG nova.compute.manager [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 557.515134] env[63531]: DEBUG nova.virt.hardware [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 557.517741] env[63531]: DEBUG nova.virt.hardware [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 557.521023] env[63531]: DEBUG nova.virt.hardware [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 557.521023] env[63531]: DEBUG nova.virt.hardware [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 557.521023] env[63531]: DEBUG nova.virt.hardware [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 557.521023] env[63531]: DEBUG nova.virt.hardware [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 557.521023] env[63531]: DEBUG nova.virt.hardware [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 557.521023] env[63531]: DEBUG nova.virt.hardware [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 557.521384] env[63531]: DEBUG nova.virt.hardware [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 557.521384] env[63531]: DEBUG nova.virt.hardware [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 557.521384] env[63531]: DEBUG nova.virt.hardware [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 557.521384] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f8b3931-b46b-4cf4-baa6-b40015b4d149 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.536505] env[63531]: DEBUG oslo_vmware.api [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': task-1117462, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 557.538155] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05f88fa9-61ac-4bf8-bfbf-3e4c622195d0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.559640] env[63531]: ERROR nova.compute.manager [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b832b20c-e3b8-40a9-a4cf-91f8c961fd94, please check neutron logs for more information. [ 557.559640] env[63531]: ERROR nova.compute.manager Traceback (most recent call last): [ 557.559640] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 557.559640] env[63531]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 557.559640] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 557.559640] env[63531]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 557.559640] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 557.559640] env[63531]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 557.559640] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 557.559640] env[63531]: ERROR nova.compute.manager self.force_reraise() [ 557.559640] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 557.559640] env[63531]: ERROR nova.compute.manager raise self.value [ 557.559640] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 557.559640] env[63531]: ERROR nova.compute.manager updated_port = self._update_port( [ 557.559640] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 557.559640] env[63531]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 557.560156] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 557.560156] env[63531]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 557.560156] env[63531]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b832b20c-e3b8-40a9-a4cf-91f8c961fd94, please check neutron logs for more information. [ 557.560156] env[63531]: ERROR nova.compute.manager [ 557.560156] env[63531]: Traceback (most recent call last): [ 557.560156] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 557.560156] env[63531]: listener.cb(fileno) [ 557.560156] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 557.560156] env[63531]: result = function(*args, **kwargs) [ 557.560156] env[63531]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 557.560156] env[63531]: return func(*args, **kwargs) [ 557.560156] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 557.560156] env[63531]: raise e [ 557.560156] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 557.560156] env[63531]: nwinfo = self.network_api.allocate_for_instance( [ 557.560156] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 557.560156] env[63531]: created_port_ids = self._update_ports_for_instance( [ 557.560156] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 557.560156] env[63531]: with excutils.save_and_reraise_exception(): [ 557.560156] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 557.560156] env[63531]: self.force_reraise() [ 557.560156] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 557.560156] env[63531]: raise self.value [ 557.560156] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 557.560156] env[63531]: updated_port = self._update_port( [ 557.560156] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 557.560156] env[63531]: _ensure_no_port_binding_failure(port) [ 557.560156] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 557.560156] env[63531]: raise exception.PortBindingFailed(port_id=port['id']) [ 557.560901] env[63531]: nova.exception.PortBindingFailed: Binding failed for port b832b20c-e3b8-40a9-a4cf-91f8c961fd94, please check neutron logs for more information. [ 557.560901] env[63531]: Removing descriptor: 17 [ 557.560901] env[63531]: ERROR nova.compute.manager [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b832b20c-e3b8-40a9-a4cf-91f8c961fd94, please check neutron logs for more information. [ 557.560901] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] Traceback (most recent call last): [ 557.560901] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 557.560901] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] yield resources [ 557.560901] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 557.560901] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] self.driver.spawn(context, instance, image_meta, [ 557.560901] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 557.560901] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 557.560901] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 557.560901] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] vm_ref = self.build_virtual_machine(instance, [ 557.561291] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 557.561291] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] vif_infos = vmwarevif.get_vif_info(self._session, [ 557.561291] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 557.561291] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] for vif in network_info: [ 557.561291] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 557.561291] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] return self._sync_wrapper(fn, *args, **kwargs) [ 557.561291] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 557.561291] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] self.wait() [ 557.561291] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 557.561291] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] self[:] = self._gt.wait() [ 557.561291] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 557.561291] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] return self._exit_event.wait() [ 557.561291] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 557.561644] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] result = hub.switch() [ 557.561644] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 557.561644] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] return self.greenlet.switch() [ 557.561644] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 557.561644] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] result = function(*args, **kwargs) [ 557.561644] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 557.561644] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] return func(*args, **kwargs) [ 557.561644] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 557.561644] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] raise e [ 557.561644] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 557.561644] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] nwinfo = self.network_api.allocate_for_instance( [ 557.561644] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 557.561644] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] created_port_ids = self._update_ports_for_instance( [ 557.562085] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 557.562085] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] with excutils.save_and_reraise_exception(): [ 557.562085] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 557.562085] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] self.force_reraise() [ 557.562085] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 557.562085] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] raise self.value [ 557.562085] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 557.562085] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] updated_port = self._update_port( [ 557.562085] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 557.562085] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] _ensure_no_port_binding_failure(port) [ 557.562085] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 557.562085] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] raise exception.PortBindingFailed(port_id=port['id']) [ 557.562426] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] nova.exception.PortBindingFailed: Binding failed for port b832b20c-e3b8-40a9-a4cf-91f8c961fd94, please check neutron logs for more information. [ 557.562426] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] [ 557.562426] env[63531]: INFO nova.compute.manager [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] Terminating instance [ 557.563190] env[63531]: DEBUG oslo_concurrency.lockutils [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] Acquiring lock "refresh_cache-c552e749-33ec-4093-be8c-ed3784bdc2b5" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 557.563337] env[63531]: DEBUG oslo_concurrency.lockutils [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] Acquired lock "refresh_cache-c552e749-33ec-4093-be8c-ed3784bdc2b5" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 557.563504] env[63531]: DEBUG nova.network.neutron [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 557.746307] env[63531]: DEBUG nova.compute.manager [req-13bfe199-a12d-4b01-be94-f8e2487c981b req-1b65ab9f-63de-4c95-b6f9-ff1f1585db85 service nova] [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] Received event network-changed-b832b20c-e3b8-40a9-a4cf-91f8c961fd94 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 557.746479] env[63531]: DEBUG nova.compute.manager [req-13bfe199-a12d-4b01-be94-f8e2487c981b req-1b65ab9f-63de-4c95-b6f9-ff1f1585db85 service nova] [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] Refreshing instance network info cache due to event network-changed-b832b20c-e3b8-40a9-a4cf-91f8c961fd94. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 557.746615] env[63531]: DEBUG oslo_concurrency.lockutils [req-13bfe199-a12d-4b01-be94-f8e2487c981b req-1b65ab9f-63de-4c95-b6f9-ff1f1585db85 service nova] Acquiring lock "refresh_cache-c552e749-33ec-4093-be8c-ed3784bdc2b5" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 557.787192] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 557.787550] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 557.787794] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Starting heal instance info cache {{(pid=63531) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 557.787870] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Rebuilding the list of instances to heal {{(pid=63531) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 557.841220] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.383s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 557.841854] env[63531]: DEBUG nova.compute.manager [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 557.845364] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.977s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 557.846963] env[63531]: INFO nova.compute.claims [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 557.949741] env[63531]: DEBUG oslo_concurrency.lockutils [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Acquiring lock "9fa373b3-8e15-4d26-82be-7dda51bc6cf6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 557.950023] env[63531]: DEBUG oslo_concurrency.lockutils [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Lock "9fa373b3-8e15-4d26-82be-7dda51bc6cf6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 558.031745] env[63531]: DEBUG oslo_vmware.api [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': task-1117462, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.677385} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 558.032539] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Copied Virtual Disk [datastore1] vmware_temp/f265b2d4-8bb5-4203-8958-00950d4e1576/9d550399-588c-4bd5-8436-eba9b523bd13/tmp-sparse.vmdk to [datastore1] vmware_temp/f265b2d4-8bb5-4203-8958-00950d4e1576/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 558.032539] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Deleting the datastore file [datastore1] vmware_temp/f265b2d4-8bb5-4203-8958-00950d4e1576/9d550399-588c-4bd5-8436-eba9b523bd13/tmp-sparse.vmdk {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 558.032984] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6f8b8b8e-996b-4142-8485-f53021f2be21 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.042195] env[63531]: DEBUG oslo_vmware.api [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Waiting for the task: (returnval){ [ 558.042195] env[63531]: value = "task-1117463" [ 558.042195] env[63531]: _type = "Task" [ 558.042195] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 558.052669] env[63531]: DEBUG oslo_vmware.api [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': task-1117463, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 558.148114] env[63531]: DEBUG nova.network.neutron [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 558.303223] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] Skipping network cache update for instance because it is Building. {{(pid=63531) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 558.307216] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] Skipping network cache update for instance because it is Building. {{(pid=63531) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 558.307216] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Skipping network cache update for instance because it is Building. {{(pid=63531) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 558.307216] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] Skipping network cache update for instance because it is Building. {{(pid=63531) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 558.308228] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] Skipping network cache update for instance because it is Building. {{(pid=63531) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 558.308812] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] Skipping network cache update for instance because it is Building. {{(pid=63531) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 558.312275] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] Skipping network cache update for instance because it is Building. {{(pid=63531) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 558.312275] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Didn't find any instances for network info cache update. {{(pid=63531) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 558.312275] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 558.316421] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 558.316703] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 558.317847] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 558.317847] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 558.317847] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 558.317847] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63531) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 558.317847] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager.update_available_resource {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 558.351408] env[63531]: DEBUG nova.compute.utils [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 558.352741] env[63531]: DEBUG nova.compute.manager [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 558.352912] env[63531]: DEBUG nova.network.neutron [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 558.473259] env[63531]: DEBUG nova.policy [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3be823dc41cb483f9caa323f3a535fb4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '26fad73663a34fc5b2ce6a2a975a6e7f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 558.554529] env[63531]: DEBUG oslo_vmware.api [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': task-1117463, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.023072} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 558.556076] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 558.556437] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Moving file from [datastore1] vmware_temp/f265b2d4-8bb5-4203-8958-00950d4e1576/9d550399-588c-4bd5-8436-eba9b523bd13 to [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13. {{(pid=63531) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 558.556820] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-7b213cba-c11f-4223-b228-af479f374271 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.566730] env[63531]: DEBUG oslo_vmware.api [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Waiting for the task: (returnval){ [ 558.566730] env[63531]: value = "task-1117464" [ 558.566730] env[63531]: _type = "Task" [ 558.566730] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 558.579061] env[63531]: DEBUG oslo_vmware.api [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': task-1117464, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 558.600734] env[63531]: DEBUG nova.network.neutron [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 558.823397] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 558.857456] env[63531]: DEBUG nova.compute.manager [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 559.083323] env[63531]: DEBUG oslo_vmware.api [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': task-1117464, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.031298} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 559.083629] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] File moved {{(pid=63531) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 559.083827] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Cleaning up location [datastore1] vmware_temp/f265b2d4-8bb5-4203-8958-00950d4e1576 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 559.090027] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Deleting the datastore file [datastore1] vmware_temp/f265b2d4-8bb5-4203-8958-00950d4e1576 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 559.090175] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-55b35b0c-5690-4e98-9442-5f2060845c33 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.097830] env[63531]: DEBUG oslo_vmware.api [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Waiting for the task: (returnval){ [ 559.097830] env[63531]: value = "task-1117465" [ 559.097830] env[63531]: _type = "Task" [ 559.097830] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 559.111152] env[63531]: DEBUG oslo_concurrency.lockutils [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] Releasing lock "refresh_cache-c552e749-33ec-4093-be8c-ed3784bdc2b5" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 559.111152] env[63531]: DEBUG nova.compute.manager [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 559.111152] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 559.117312] env[63531]: DEBUG oslo_concurrency.lockutils [req-13bfe199-a12d-4b01-be94-f8e2487c981b req-1b65ab9f-63de-4c95-b6f9-ff1f1585db85 service nova] Acquired lock "refresh_cache-c552e749-33ec-4093-be8c-ed3784bdc2b5" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 559.117540] env[63531]: DEBUG nova.network.neutron [req-13bfe199-a12d-4b01-be94-f8e2487c981b req-1b65ab9f-63de-4c95-b6f9-ff1f1585db85 service nova] [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] Refreshing network info cache for port b832b20c-e3b8-40a9-a4cf-91f8c961fd94 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 559.118753] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-23b3a639-3795-4aa4-b1af-57da813372d1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.129275] env[63531]: DEBUG oslo_vmware.api [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': task-1117465, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.026452} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 559.130056] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 559.130823] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d4a8cc13-4ba8-4169-808d-a1c227788507 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.143022] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-605e7359-5b74-44e0-9240-5ae5c29cd658 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.155473] env[63531]: DEBUG oslo_vmware.api [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Waiting for the task: (returnval){ [ 559.155473] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]527d95cd-bf31-0d13-3e84-7b6f67d3cfc9" [ 559.155473] env[63531]: _type = "Task" [ 559.155473] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 559.157845] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4725eed3-5361-4367-8f37-30cfbda8f4c4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.169038] env[63531]: WARNING nova.virt.vmwareapi.vmops [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c552e749-33ec-4093-be8c-ed3784bdc2b5 could not be found. [ 559.169318] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 559.169509] env[63531]: INFO nova.compute.manager [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] Took 0.06 seconds to destroy the instance on the hypervisor. [ 559.169857] env[63531]: DEBUG oslo.service.loopingcall [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 559.173400] env[63531]: DEBUG nova.compute.manager [-] [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 559.173400] env[63531]: DEBUG nova.network.neutron [-] [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 559.176869] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1599d1e-c093-4973-86e4-ce5325943bb0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.218961] env[63531]: DEBUG oslo_vmware.api [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]527d95cd-bf31-0d13-3e84-7b6f67d3cfc9, 'name': SearchDatastore_Task, 'duration_secs': 0.009075} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 559.219815] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2d18611-2c30-4cdb-9b85-6d341da53960 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.222973] env[63531]: DEBUG oslo_concurrency.lockutils [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 559.223319] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 4410b013-868a-41e7-9dfb-584fb2d9d142/4410b013-868a-41e7-9dfb-584fb2d9d142.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 559.223736] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c90683f2-8a3d-4e26-890b-0856a9a3bac6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.235081] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f044781-23a3-425e-9b19-59011291880c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.244736] env[63531]: DEBUG oslo_vmware.api [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Waiting for the task: (returnval){ [ 559.244736] env[63531]: value = "task-1117466" [ 559.244736] env[63531]: _type = "Task" [ 559.244736] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 559.256909] env[63531]: DEBUG nova.compute.provider_tree [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 559.262795] env[63531]: DEBUG oslo_vmware.api [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': task-1117466, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 559.289842] env[63531]: DEBUG nova.network.neutron [-] [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 559.395520] env[63531]: DEBUG nova.network.neutron [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] Successfully created port: e006227b-9069-49b7-a244-083d6aed19d7 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 559.686780] env[63531]: DEBUG nova.network.neutron [req-13bfe199-a12d-4b01-be94-f8e2487c981b req-1b65ab9f-63de-4c95-b6f9-ff1f1585db85 service nova] [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 559.759750] env[63531]: DEBUG oslo_vmware.api [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': task-1117466, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.478668} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 559.760874] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 4410b013-868a-41e7-9dfb-584fb2d9d142/4410b013-868a-41e7-9dfb-584fb2d9d142.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 559.760874] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 559.761030] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9730619f-e9d3-4546-ac87-860e080a0814 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.767048] env[63531]: DEBUG nova.scheduler.client.report [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 559.772129] env[63531]: DEBUG oslo_vmware.api [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Waiting for the task: (returnval){ [ 559.772129] env[63531]: value = "task-1117467" [ 559.772129] env[63531]: _type = "Task" [ 559.772129] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 559.785116] env[63531]: DEBUG oslo_vmware.api [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': task-1117467, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 559.792462] env[63531]: DEBUG nova.network.neutron [-] [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 559.873143] env[63531]: DEBUG nova.compute.manager [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 559.906017] env[63531]: DEBUG nova.virt.hardware [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 559.906017] env[63531]: DEBUG nova.virt.hardware [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 559.906017] env[63531]: DEBUG nova.virt.hardware [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 559.906281] env[63531]: DEBUG nova.virt.hardware [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 559.906281] env[63531]: DEBUG nova.virt.hardware [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 559.906281] env[63531]: DEBUG nova.virt.hardware [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 559.906281] env[63531]: DEBUG nova.virt.hardware [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 559.906281] env[63531]: DEBUG nova.virt.hardware [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 559.906458] env[63531]: DEBUG nova.virt.hardware [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 559.906458] env[63531]: DEBUG nova.virt.hardware [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 559.906458] env[63531]: DEBUG nova.virt.hardware [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 559.907617] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38f37476-935d-4acd-a471-7fa0772c2d17 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.917251] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f2991e0-aad1-46ca-a827-48314549c46e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.096137] env[63531]: DEBUG nova.network.neutron [req-13bfe199-a12d-4b01-be94-f8e2487c981b req-1b65ab9f-63de-4c95-b6f9-ff1f1585db85 service nova] [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 560.185447] env[63531]: DEBUG nova.compute.manager [req-a21b6c98-f111-4405-9490-d6cd624415a2 req-2577f155-3998-4158-a019-d745ef0c968e service nova] [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] Received event network-changed-73d60e7d-c4d2-475d-aeed-b12ea040e3df {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 560.185674] env[63531]: DEBUG nova.compute.manager [req-a21b6c98-f111-4405-9490-d6cd624415a2 req-2577f155-3998-4158-a019-d745ef0c968e service nova] [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] Refreshing instance network info cache due to event network-changed-73d60e7d-c4d2-475d-aeed-b12ea040e3df. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 560.185856] env[63531]: DEBUG oslo_concurrency.lockutils [req-a21b6c98-f111-4405-9490-d6cd624415a2 req-2577f155-3998-4158-a019-d745ef0c968e service nova] Acquiring lock "refresh_cache-d1636eb1-b862-47c9-8729-7a9d475ace01" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 560.186050] env[63531]: DEBUG oslo_concurrency.lockutils [req-a21b6c98-f111-4405-9490-d6cd624415a2 req-2577f155-3998-4158-a019-d745ef0c968e service nova] Acquired lock "refresh_cache-d1636eb1-b862-47c9-8729-7a9d475ace01" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 560.186227] env[63531]: DEBUG nova.network.neutron [req-a21b6c98-f111-4405-9490-d6cd624415a2 req-2577f155-3998-4158-a019-d745ef0c968e service nova] [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] Refreshing network info cache for port 73d60e7d-c4d2-475d-aeed-b12ea040e3df {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 560.269705] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.424s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 560.270309] env[63531]: DEBUG nova.compute.manager [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 560.272909] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.258s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.274116] env[63531]: INFO nova.compute.claims [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] [instance: 5b86b63a-f6c1-4983-a3cc-3cde7a10d17b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 560.293108] env[63531]: DEBUG oslo_vmware.api [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': task-1117467, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060416} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 560.293466] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 560.294416] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5b053fe-54ec-4ab3-b01d-c050abd91da2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.299032] env[63531]: INFO nova.compute.manager [-] [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] Took 1.12 seconds to deallocate network for instance. [ 560.301323] env[63531]: DEBUG nova.compute.claims [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] Aborting claim: {{(pid=63531) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 560.302027] env[63531]: DEBUG oslo_concurrency.lockutils [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.322408] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Reconfiguring VM instance instance-00000003 to attach disk [datastore1] 4410b013-868a-41e7-9dfb-584fb2d9d142/4410b013-868a-41e7-9dfb-584fb2d9d142.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 560.323610] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0c0e6777-2f22-4440-a7db-cbb4ec756ba8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.346991] env[63531]: DEBUG oslo_vmware.api [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Waiting for the task: (returnval){ [ 560.346991] env[63531]: value = "task-1117468" [ 560.346991] env[63531]: _type = "Task" [ 560.346991] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 560.356889] env[63531]: DEBUG oslo_vmware.api [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': task-1117468, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 560.477319] env[63531]: ERROR nova.compute.manager [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 73d60e7d-c4d2-475d-aeed-b12ea040e3df, please check neutron logs for more information. [ 560.477319] env[63531]: ERROR nova.compute.manager Traceback (most recent call last): [ 560.477319] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 560.477319] env[63531]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 560.477319] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 560.477319] env[63531]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 560.477319] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 560.477319] env[63531]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 560.477319] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 560.477319] env[63531]: ERROR nova.compute.manager self.force_reraise() [ 560.477319] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 560.477319] env[63531]: ERROR nova.compute.manager raise self.value [ 560.477319] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 560.477319] env[63531]: ERROR nova.compute.manager updated_port = self._update_port( [ 560.477319] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 560.477319] env[63531]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 560.478625] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 560.478625] env[63531]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 560.478625] env[63531]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 73d60e7d-c4d2-475d-aeed-b12ea040e3df, please check neutron logs for more information. [ 560.478625] env[63531]: ERROR nova.compute.manager [ 560.478625] env[63531]: Traceback (most recent call last): [ 560.478625] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 560.478625] env[63531]: listener.cb(fileno) [ 560.478625] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 560.478625] env[63531]: result = function(*args, **kwargs) [ 560.478625] env[63531]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 560.478625] env[63531]: return func(*args, **kwargs) [ 560.478625] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 560.478625] env[63531]: raise e [ 560.478625] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 560.478625] env[63531]: nwinfo = self.network_api.allocate_for_instance( [ 560.478625] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 560.478625] env[63531]: created_port_ids = self._update_ports_for_instance( [ 560.478625] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 560.478625] env[63531]: with excutils.save_and_reraise_exception(): [ 560.478625] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 560.478625] env[63531]: self.force_reraise() [ 560.478625] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 560.478625] env[63531]: raise self.value [ 560.478625] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 560.478625] env[63531]: updated_port = self._update_port( [ 560.478625] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 560.478625] env[63531]: _ensure_no_port_binding_failure(port) [ 560.478625] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 560.478625] env[63531]: raise exception.PortBindingFailed(port_id=port['id']) [ 560.480012] env[63531]: nova.exception.PortBindingFailed: Binding failed for port 73d60e7d-c4d2-475d-aeed-b12ea040e3df, please check neutron logs for more information. [ 560.480012] env[63531]: Removing descriptor: 15 [ 560.480012] env[63531]: ERROR nova.compute.manager [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 73d60e7d-c4d2-475d-aeed-b12ea040e3df, please check neutron logs for more information. [ 560.480012] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] Traceback (most recent call last): [ 560.480012] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 560.480012] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] yield resources [ 560.480012] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 560.480012] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] self.driver.spawn(context, instance, image_meta, [ 560.480012] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 560.480012] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] self._vmops.spawn(context, instance, image_meta, injected_files, [ 560.480012] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 560.480012] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] vm_ref = self.build_virtual_machine(instance, [ 560.480344] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 560.480344] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] vif_infos = vmwarevif.get_vif_info(self._session, [ 560.480344] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 560.480344] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] for vif in network_info: [ 560.480344] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 560.480344] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] return self._sync_wrapper(fn, *args, **kwargs) [ 560.480344] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 560.480344] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] self.wait() [ 560.480344] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 560.480344] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] self[:] = self._gt.wait() [ 560.480344] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 560.480344] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] return self._exit_event.wait() [ 560.480344] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 560.480769] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] result = hub.switch() [ 560.480769] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 560.480769] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] return self.greenlet.switch() [ 560.480769] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 560.480769] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] result = function(*args, **kwargs) [ 560.480769] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 560.480769] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] return func(*args, **kwargs) [ 560.480769] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 560.480769] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] raise e [ 560.480769] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 560.480769] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] nwinfo = self.network_api.allocate_for_instance( [ 560.480769] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 560.480769] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] created_port_ids = self._update_ports_for_instance( [ 560.481168] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 560.481168] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] with excutils.save_and_reraise_exception(): [ 560.481168] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 560.481168] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] self.force_reraise() [ 560.481168] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 560.481168] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] raise self.value [ 560.481168] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 560.481168] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] updated_port = self._update_port( [ 560.481168] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 560.481168] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] _ensure_no_port_binding_failure(port) [ 560.481168] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 560.481168] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] raise exception.PortBindingFailed(port_id=port['id']) [ 560.481598] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] nova.exception.PortBindingFailed: Binding failed for port 73d60e7d-c4d2-475d-aeed-b12ea040e3df, please check neutron logs for more information. [ 560.481598] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] [ 560.481598] env[63531]: INFO nova.compute.manager [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] Terminating instance [ 560.481598] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Acquiring lock "refresh_cache-d1636eb1-b862-47c9-8729-7a9d475ace01" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 560.602237] env[63531]: DEBUG oslo_concurrency.lockutils [req-13bfe199-a12d-4b01-be94-f8e2487c981b req-1b65ab9f-63de-4c95-b6f9-ff1f1585db85 service nova] Releasing lock "refresh_cache-c552e749-33ec-4093-be8c-ed3784bdc2b5" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 560.749087] env[63531]: DEBUG nova.network.neutron [req-a21b6c98-f111-4405-9490-d6cd624415a2 req-2577f155-3998-4158-a019-d745ef0c968e service nova] [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 560.778374] env[63531]: DEBUG nova.compute.utils [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 560.779768] env[63531]: DEBUG nova.compute.manager [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 560.779940] env[63531]: DEBUG nova.network.neutron [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 560.793744] env[63531]: DEBUG nova.compute.manager [req-9396deea-49d8-4a6d-a062-e194fd445df5 req-e34becf3-7ebf-485e-995e-61057b64ed6c service nova] [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] Received event network-vif-deleted-b832b20c-e3b8-40a9-a4cf-91f8c961fd94 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 560.859542] env[63531]: DEBUG oslo_vmware.api [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': task-1117468, 'name': ReconfigVM_Task, 'duration_secs': 0.271912} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 560.860108] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Reconfigured VM instance instance-00000003 to attach disk [datastore1] 4410b013-868a-41e7-9dfb-584fb2d9d142/4410b013-868a-41e7-9dfb-584fb2d9d142.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 560.860950] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5f3cb1fd-baa7-491d-9cb1-d8de6ae1fc4f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.868271] env[63531]: DEBUG oslo_vmware.api [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Waiting for the task: (returnval){ [ 560.868271] env[63531]: value = "task-1117469" [ 560.868271] env[63531]: _type = "Task" [ 560.868271] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 560.879025] env[63531]: DEBUG oslo_vmware.api [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': task-1117469, 'name': Rename_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 561.015805] env[63531]: DEBUG nova.policy [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '85b959a54bf94b0e9f7f917ea1382be7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c5d85705a4424e8ba2cf87d5011fb952', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 561.024102] env[63531]: DEBUG nova.network.neutron [req-a21b6c98-f111-4405-9490-d6cd624415a2 req-2577f155-3998-4158-a019-d745ef0c968e service nova] [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 561.287080] env[63531]: DEBUG nova.compute.manager [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 561.384834] env[63531]: DEBUG oslo_vmware.api [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': task-1117469, 'name': Rename_Task, 'duration_secs': 0.132704} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 561.385441] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 561.385441] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-60dd56b3-4c40-4bd7-8135-ceaee9a6fe5c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.395348] env[63531]: DEBUG oslo_vmware.api [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Waiting for the task: (returnval){ [ 561.395348] env[63531]: value = "task-1117470" [ 561.395348] env[63531]: _type = "Task" [ 561.395348] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 561.406040] env[63531]: DEBUG oslo_vmware.api [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': task-1117470, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 561.524505] env[63531]: DEBUG oslo_concurrency.lockutils [req-a21b6c98-f111-4405-9490-d6cd624415a2 req-2577f155-3998-4158-a019-d745ef0c968e service nova] Releasing lock "refresh_cache-d1636eb1-b862-47c9-8729-7a9d475ace01" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 561.526783] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Acquired lock "refresh_cache-d1636eb1-b862-47c9-8729-7a9d475ace01" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 561.526783] env[63531]: DEBUG nova.network.neutron [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 561.556569] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8521cc98-8fc9-4bab-8d48-ae4740bcba98 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.567692] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b065db2c-4f49-4aa9-b313-238cee22d393 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.602851] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71b59b80-63e8-4069-94be-f025c55750dc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.610985] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab0c9354-3f37-45ef-a078-9ac24d55b778 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.625576] env[63531]: DEBUG nova.compute.provider_tree [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 561.905695] env[63531]: DEBUG oslo_vmware.api [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': task-1117470, 'name': PowerOnVM_Task, 'duration_secs': 0.426769} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 561.907057] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 561.907057] env[63531]: INFO nova.compute.manager [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Took 11.40 seconds to spawn the instance on the hypervisor. [ 561.907057] env[63531]: DEBUG nova.compute.manager [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 561.907443] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-562045e3-adec-4ad0-a3ad-e5a3e626cc52 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.073269] env[63531]: DEBUG nova.network.neutron [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 562.128883] env[63531]: DEBUG nova.scheduler.client.report [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 562.284147] env[63531]: DEBUG nova.network.neutron [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 562.299665] env[63531]: DEBUG nova.compute.manager [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 562.336616] env[63531]: DEBUG nova.virt.hardware [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 562.336616] env[63531]: DEBUG nova.virt.hardware [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 562.336745] env[63531]: DEBUG nova.virt.hardware [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 562.337665] env[63531]: DEBUG nova.virt.hardware [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 562.337665] env[63531]: DEBUG nova.virt.hardware [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 562.337665] env[63531]: DEBUG nova.virt.hardware [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 562.337665] env[63531]: DEBUG nova.virt.hardware [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 562.337665] env[63531]: DEBUG nova.virt.hardware [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 562.337914] env[63531]: DEBUG nova.virt.hardware [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 562.338036] env[63531]: DEBUG nova.virt.hardware [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 562.338257] env[63531]: DEBUG nova.virt.hardware [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 562.339161] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11d4ee5e-41f7-451d-855c-e1ab1f4982fc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.347425] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22eee63b-c00c-4525-bb89-5e0b126d705c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.430073] env[63531]: INFO nova.compute.manager [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Took 18.44 seconds to build instance. [ 562.636819] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.362s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 562.636819] env[63531]: DEBUG nova.compute.manager [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] [instance: 5b86b63a-f6c1-4983-a3cc-3cde7a10d17b] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 562.641053] env[63531]: DEBUG oslo_concurrency.lockutils [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 9.338s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 562.684438] env[63531]: DEBUG nova.network.neutron [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] Successfully created port: 60020572-c02a-4101-9cec-88afc59a70d9 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 562.794072] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Releasing lock "refresh_cache-d1636eb1-b862-47c9-8729-7a9d475ace01" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 562.794072] env[63531]: DEBUG nova.compute.manager [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 562.794072] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 562.794072] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-76013f85-8052-481b-9d0e-aafd005f1970 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.809413] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0215801e-e279-4099-9956-b151a2cc03e4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.841286] env[63531]: WARNING nova.virt.vmwareapi.vmops [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d1636eb1-b862-47c9-8729-7a9d475ace01 could not be found. [ 562.841286] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 562.841286] env[63531]: INFO nova.compute.manager [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] Took 0.05 seconds to destroy the instance on the hypervisor. [ 562.841286] env[63531]: DEBUG oslo.service.loopingcall [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 562.842648] env[63531]: DEBUG nova.compute.manager [-] [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 562.842648] env[63531]: DEBUG nova.network.neutron [-] [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 562.905619] env[63531]: DEBUG nova.network.neutron [-] [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 562.936084] env[63531]: DEBUG oslo_concurrency.lockutils [None req-73c396be-b130-477b-9ab0-a35c51f8d7ac tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Lock "4410b013-868a-41e7-9dfb-584fb2d9d142" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.960s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 563.154016] env[63531]: DEBUG nova.compute.utils [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 563.161098] env[63531]: DEBUG nova.compute.manager [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] [instance: 5b86b63a-f6c1-4983-a3cc-3cde7a10d17b] Not allocating networking since 'none' was specified. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 563.308475] env[63531]: DEBUG nova.compute.manager [req-9b314f4b-8070-4dc2-bdd6-7657fe2c6df7 req-04f42b6c-9056-4ac4-8c11-87acb506d1fa service nova] [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] Received event network-vif-deleted-73d60e7d-c4d2-475d-aeed-b12ea040e3df {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 563.369254] env[63531]: ERROR nova.compute.manager [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e70dcd41-edfa-4975-a540-6e7d7bbd6280, please check neutron logs for more information. [ 563.369254] env[63531]: ERROR nova.compute.manager Traceback (most recent call last): [ 563.369254] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 563.369254] env[63531]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 563.369254] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 563.369254] env[63531]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 563.369254] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 563.369254] env[63531]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 563.369254] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 563.369254] env[63531]: ERROR nova.compute.manager self.force_reraise() [ 563.369254] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 563.369254] env[63531]: ERROR nova.compute.manager raise self.value [ 563.369254] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 563.369254] env[63531]: ERROR nova.compute.manager updated_port = self._update_port( [ 563.369254] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 563.369254] env[63531]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 563.369893] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 563.369893] env[63531]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 563.369893] env[63531]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e70dcd41-edfa-4975-a540-6e7d7bbd6280, please check neutron logs for more information. [ 563.369893] env[63531]: ERROR nova.compute.manager [ 563.369893] env[63531]: Traceback (most recent call last): [ 563.369893] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 563.369893] env[63531]: listener.cb(fileno) [ 563.369893] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 563.369893] env[63531]: result = function(*args, **kwargs) [ 563.369893] env[63531]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 563.369893] env[63531]: return func(*args, **kwargs) [ 563.369893] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 563.369893] env[63531]: raise e [ 563.369893] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 563.369893] env[63531]: nwinfo = self.network_api.allocate_for_instance( [ 563.369893] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 563.369893] env[63531]: created_port_ids = self._update_ports_for_instance( [ 563.369893] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 563.369893] env[63531]: with excutils.save_and_reraise_exception(): [ 563.369893] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 563.369893] env[63531]: self.force_reraise() [ 563.369893] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 563.369893] env[63531]: raise self.value [ 563.369893] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 563.369893] env[63531]: updated_port = self._update_port( [ 563.369893] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 563.369893] env[63531]: _ensure_no_port_binding_failure(port) [ 563.369893] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 563.369893] env[63531]: raise exception.PortBindingFailed(port_id=port['id']) [ 563.372400] env[63531]: nova.exception.PortBindingFailed: Binding failed for port e70dcd41-edfa-4975-a540-6e7d7bbd6280, please check neutron logs for more information. [ 563.372400] env[63531]: Removing descriptor: 20 [ 563.372400] env[63531]: ERROR nova.compute.manager [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e70dcd41-edfa-4975-a540-6e7d7bbd6280, please check neutron logs for more information. [ 563.372400] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] Traceback (most recent call last): [ 563.372400] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 563.372400] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] yield resources [ 563.372400] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 563.372400] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] self.driver.spawn(context, instance, image_meta, [ 563.372400] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 563.372400] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] self._vmops.spawn(context, instance, image_meta, injected_files, [ 563.372400] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 563.372400] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] vm_ref = self.build_virtual_machine(instance, [ 563.372842] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 563.372842] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] vif_infos = vmwarevif.get_vif_info(self._session, [ 563.372842] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 563.372842] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] for vif in network_info: [ 563.372842] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 563.372842] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] return self._sync_wrapper(fn, *args, **kwargs) [ 563.372842] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 563.372842] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] self.wait() [ 563.372842] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 563.372842] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] self[:] = self._gt.wait() [ 563.372842] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 563.372842] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] return self._exit_event.wait() [ 563.372842] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 563.373284] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] result = hub.switch() [ 563.373284] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 563.373284] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] return self.greenlet.switch() [ 563.373284] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 563.373284] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] result = function(*args, **kwargs) [ 563.373284] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 563.373284] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] return func(*args, **kwargs) [ 563.373284] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 563.373284] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] raise e [ 563.373284] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 563.373284] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] nwinfo = self.network_api.allocate_for_instance( [ 563.373284] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 563.373284] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] created_port_ids = self._update_ports_for_instance( [ 563.373677] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 563.373677] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] with excutils.save_and_reraise_exception(): [ 563.373677] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 563.373677] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] self.force_reraise() [ 563.373677] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 563.373677] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] raise self.value [ 563.373677] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 563.373677] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] updated_port = self._update_port( [ 563.373677] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 563.373677] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] _ensure_no_port_binding_failure(port) [ 563.373677] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 563.373677] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] raise exception.PortBindingFailed(port_id=port['id']) [ 563.374675] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] nova.exception.PortBindingFailed: Binding failed for port e70dcd41-edfa-4975-a540-6e7d7bbd6280, please check neutron logs for more information. [ 563.374675] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] [ 563.374675] env[63531]: INFO nova.compute.manager [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] Terminating instance [ 563.374675] env[63531]: DEBUG oslo_concurrency.lockutils [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Acquiring lock "refresh_cache-6e963227-64f7-414d-88ec-cee5d5770bba" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 563.374675] env[63531]: DEBUG oslo_concurrency.lockutils [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Acquired lock "refresh_cache-6e963227-64f7-414d-88ec-cee5d5770bba" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 563.374675] env[63531]: DEBUG nova.network.neutron [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 563.409161] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-472bf0f3-c48a-446c-bc6b-178bb0ea6af3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.412950] env[63531]: DEBUG nova.network.neutron [-] [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 563.425150] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7afba86f-6360-41b5-8196-5473cee06313 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.459814] env[63531]: DEBUG nova.compute.manager [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 563.463753] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b14083e-bebf-4743-a203-66022f075dcd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.474428] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c041e750-2c01-48fd-b197-aa3a9ff6694b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.497357] env[63531]: DEBUG nova.compute.provider_tree [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 563.664325] env[63531]: DEBUG nova.compute.manager [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] [instance: 5b86b63a-f6c1-4983-a3cc-3cde7a10d17b] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 563.775793] env[63531]: ERROR nova.compute.manager [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e006227b-9069-49b7-a244-083d6aed19d7, please check neutron logs for more information. [ 563.775793] env[63531]: ERROR nova.compute.manager Traceback (most recent call last): [ 563.775793] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 563.775793] env[63531]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 563.775793] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 563.775793] env[63531]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 563.775793] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 563.775793] env[63531]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 563.775793] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 563.775793] env[63531]: ERROR nova.compute.manager self.force_reraise() [ 563.775793] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 563.775793] env[63531]: ERROR nova.compute.manager raise self.value [ 563.775793] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 563.775793] env[63531]: ERROR nova.compute.manager updated_port = self._update_port( [ 563.775793] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 563.775793] env[63531]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 563.776391] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 563.776391] env[63531]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 563.776391] env[63531]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e006227b-9069-49b7-a244-083d6aed19d7, please check neutron logs for more information. [ 563.776391] env[63531]: ERROR nova.compute.manager [ 563.776391] env[63531]: Traceback (most recent call last): [ 563.776391] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 563.776391] env[63531]: listener.cb(fileno) [ 563.776391] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 563.776391] env[63531]: result = function(*args, **kwargs) [ 563.776391] env[63531]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 563.776391] env[63531]: return func(*args, **kwargs) [ 563.776391] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 563.776391] env[63531]: raise e [ 563.776391] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 563.776391] env[63531]: nwinfo = self.network_api.allocate_for_instance( [ 563.776391] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 563.776391] env[63531]: created_port_ids = self._update_ports_for_instance( [ 563.776391] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 563.776391] env[63531]: with excutils.save_and_reraise_exception(): [ 563.776391] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 563.776391] env[63531]: self.force_reraise() [ 563.776391] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 563.776391] env[63531]: raise self.value [ 563.776391] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 563.776391] env[63531]: updated_port = self._update_port( [ 563.776391] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 563.776391] env[63531]: _ensure_no_port_binding_failure(port) [ 563.776391] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 563.776391] env[63531]: raise exception.PortBindingFailed(port_id=port['id']) [ 563.777155] env[63531]: nova.exception.PortBindingFailed: Binding failed for port e006227b-9069-49b7-a244-083d6aed19d7, please check neutron logs for more information. [ 563.777155] env[63531]: Removing descriptor: 14 [ 563.777769] env[63531]: ERROR nova.compute.manager [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e006227b-9069-49b7-a244-083d6aed19d7, please check neutron logs for more information. [ 563.777769] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] Traceback (most recent call last): [ 563.777769] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 563.777769] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] yield resources [ 563.777769] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 563.777769] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] self.driver.spawn(context, instance, image_meta, [ 563.777769] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 563.777769] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] self._vmops.spawn(context, instance, image_meta, injected_files, [ 563.777769] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 563.777769] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] vm_ref = self.build_virtual_machine(instance, [ 563.777769] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 563.778291] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] vif_infos = vmwarevif.get_vif_info(self._session, [ 563.778291] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 563.778291] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] for vif in network_info: [ 563.778291] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 563.778291] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] return self._sync_wrapper(fn, *args, **kwargs) [ 563.778291] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 563.778291] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] self.wait() [ 563.778291] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 563.778291] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] self[:] = self._gt.wait() [ 563.778291] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 563.778291] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] return self._exit_event.wait() [ 563.778291] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 563.778291] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] result = hub.switch() [ 563.778670] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 563.778670] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] return self.greenlet.switch() [ 563.778670] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 563.778670] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] result = function(*args, **kwargs) [ 563.778670] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 563.778670] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] return func(*args, **kwargs) [ 563.778670] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 563.778670] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] raise e [ 563.778670] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 563.778670] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] nwinfo = self.network_api.allocate_for_instance( [ 563.778670] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 563.778670] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] created_port_ids = self._update_ports_for_instance( [ 563.778670] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 563.779034] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] with excutils.save_and_reraise_exception(): [ 563.779034] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 563.779034] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] self.force_reraise() [ 563.779034] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 563.779034] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] raise self.value [ 563.779034] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 563.779034] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] updated_port = self._update_port( [ 563.779034] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 563.779034] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] _ensure_no_port_binding_failure(port) [ 563.779034] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 563.779034] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] raise exception.PortBindingFailed(port_id=port['id']) [ 563.779034] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] nova.exception.PortBindingFailed: Binding failed for port e006227b-9069-49b7-a244-083d6aed19d7, please check neutron logs for more information. [ 563.779034] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] [ 563.779409] env[63531]: INFO nova.compute.manager [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] Terminating instance [ 563.783560] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] Acquiring lock "refresh_cache-35e9c404-bad4-4f8d-8c79-fb7250abb363" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 563.783737] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] Acquired lock "refresh_cache-35e9c404-bad4-4f8d-8c79-fb7250abb363" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 563.783895] env[63531]: DEBUG nova.network.neutron [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 563.902693] env[63531]: DEBUG nova.network.neutron [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 563.916363] env[63531]: INFO nova.compute.manager [-] [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] Took 1.07 seconds to deallocate network for instance. [ 563.918869] env[63531]: DEBUG nova.compute.claims [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] Aborting claim: {{(pid=63531) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 563.919105] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.000421] env[63531]: DEBUG nova.scheduler.client.report [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 564.007539] env[63531]: DEBUG oslo_concurrency.lockutils [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.067596] env[63531]: DEBUG nova.network.neutron [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 564.243474] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] Acquiring lock "d941a593-7047-436a-b95d-bcb4850829f5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.243474] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] Lock "d941a593-7047-436a-b95d-bcb4850829f5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 564.296154] env[63531]: DEBUG nova.compute.manager [req-d43ea6b9-6932-487b-8c39-018d114936ce req-3dfa0e45-34cf-4189-bfa3-ca3ce0a0a11f service nova] [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] Received event network-changed-e70dcd41-edfa-4975-a540-6e7d7bbd6280 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 564.296154] env[63531]: DEBUG nova.compute.manager [req-d43ea6b9-6932-487b-8c39-018d114936ce req-3dfa0e45-34cf-4189-bfa3-ca3ce0a0a11f service nova] [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] Refreshing instance network info cache due to event network-changed-e70dcd41-edfa-4975-a540-6e7d7bbd6280. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 564.296154] env[63531]: DEBUG oslo_concurrency.lockutils [req-d43ea6b9-6932-487b-8c39-018d114936ce req-3dfa0e45-34cf-4189-bfa3-ca3ce0a0a11f service nova] Acquiring lock "refresh_cache-6e963227-64f7-414d-88ec-cee5d5770bba" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 564.319782] env[63531]: DEBUG nova.network.neutron [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 564.513227] env[63531]: DEBUG oslo_concurrency.lockutils [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.872s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 564.513875] env[63531]: ERROR nova.compute.manager [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 81228b1a-3a5e-426b-8ed3-d6273c990b2f, please check neutron logs for more information. [ 564.513875] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] Traceback (most recent call last): [ 564.513875] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 564.513875] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] self.driver.spawn(context, instance, image_meta, [ 564.513875] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 564.513875] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] self._vmops.spawn(context, instance, image_meta, injected_files, [ 564.513875] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 564.513875] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] vm_ref = self.build_virtual_machine(instance, [ 564.513875] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 564.513875] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] vif_infos = vmwarevif.get_vif_info(self._session, [ 564.513875] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 564.514226] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] for vif in network_info: [ 564.514226] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 564.514226] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] return self._sync_wrapper(fn, *args, **kwargs) [ 564.514226] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 564.514226] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] self.wait() [ 564.514226] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 564.514226] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] self[:] = self._gt.wait() [ 564.514226] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 564.514226] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] return self._exit_event.wait() [ 564.514226] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 564.514226] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] result = hub.switch() [ 564.514226] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 564.514226] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] return self.greenlet.switch() [ 564.514565] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 564.514565] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] result = function(*args, **kwargs) [ 564.514565] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 564.514565] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] return func(*args, **kwargs) [ 564.514565] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 564.514565] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] raise e [ 564.514565] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 564.514565] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] nwinfo = self.network_api.allocate_for_instance( [ 564.514565] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 564.514565] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] created_port_ids = self._update_ports_for_instance( [ 564.514565] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 564.514565] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] with excutils.save_and_reraise_exception(): [ 564.514565] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 564.515808] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] self.force_reraise() [ 564.515808] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 564.515808] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] raise self.value [ 564.515808] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 564.515808] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] updated_port = self._update_port( [ 564.515808] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 564.515808] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] _ensure_no_port_binding_failure(port) [ 564.515808] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 564.515808] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] raise exception.PortBindingFailed(port_id=port['id']) [ 564.515808] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] nova.exception.PortBindingFailed: Binding failed for port 81228b1a-3a5e-426b-8ed3-d6273c990b2f, please check neutron logs for more information. [ 564.515808] env[63531]: ERROR nova.compute.manager [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] [ 564.516265] env[63531]: DEBUG nova.compute.utils [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] Binding failed for port 81228b1a-3a5e-426b-8ed3-d6273c990b2f, please check neutron logs for more information. {{(pid=63531) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 564.518448] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.175s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 564.520757] env[63531]: INFO nova.compute.claims [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 564.524810] env[63531]: DEBUG nova.compute.manager [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] Build of instance ca9c41cf-b5e6-400b-9b6c-34a065722969 was re-scheduled: Binding failed for port 81228b1a-3a5e-426b-8ed3-d6273c990b2f, please check neutron logs for more information. {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 564.525313] env[63531]: DEBUG nova.compute.manager [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] Unplugging VIFs for instance {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 564.525587] env[63531]: DEBUG oslo_concurrency.lockutils [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] Acquiring lock "refresh_cache-ca9c41cf-b5e6-400b-9b6c-34a065722969" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 564.525935] env[63531]: DEBUG oslo_concurrency.lockutils [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] Acquired lock "refresh_cache-ca9c41cf-b5e6-400b-9b6c-34a065722969" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 564.526322] env[63531]: DEBUG nova.network.neutron [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 564.536277] env[63531]: DEBUG nova.network.neutron [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 564.573200] env[63531]: DEBUG oslo_concurrency.lockutils [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Releasing lock "refresh_cache-6e963227-64f7-414d-88ec-cee5d5770bba" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 564.573200] env[63531]: DEBUG nova.compute.manager [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 564.573200] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 564.573200] env[63531]: DEBUG oslo_concurrency.lockutils [req-d43ea6b9-6932-487b-8c39-018d114936ce req-3dfa0e45-34cf-4189-bfa3-ca3ce0a0a11f service nova] Acquired lock "refresh_cache-6e963227-64f7-414d-88ec-cee5d5770bba" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 564.573200] env[63531]: DEBUG nova.network.neutron [req-d43ea6b9-6932-487b-8c39-018d114936ce req-3dfa0e45-34cf-4189-bfa3-ca3ce0a0a11f service nova] [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] Refreshing network info cache for port e70dcd41-edfa-4975-a540-6e7d7bbd6280 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 564.573366] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-eaec7fb2-7878-4768-a254-abcb138d4b06 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.594523] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc8d978b-7f43-4fac-bd4e-90de8ee693e6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.621014] env[63531]: WARNING nova.virt.vmwareapi.vmops [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6e963227-64f7-414d-88ec-cee5d5770bba could not be found. [ 564.621251] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 564.621429] env[63531]: INFO nova.compute.manager [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] Took 0.05 seconds to destroy the instance on the hypervisor. [ 564.621663] env[63531]: DEBUG oslo.service.loopingcall [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 564.621896] env[63531]: DEBUG nova.compute.manager [-] [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 564.621985] env[63531]: DEBUG nova.network.neutron [-] [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 564.677034] env[63531]: DEBUG nova.network.neutron [-] [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 564.679392] env[63531]: DEBUG nova.compute.manager [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] [instance: 5b86b63a-f6c1-4983-a3cc-3cde7a10d17b] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 564.720017] env[63531]: DEBUG nova.virt.hardware [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 564.720017] env[63531]: DEBUG nova.virt.hardware [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 564.720017] env[63531]: DEBUG nova.virt.hardware [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 564.720407] env[63531]: DEBUG nova.virt.hardware [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 564.720407] env[63531]: DEBUG nova.virt.hardware [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 564.720407] env[63531]: DEBUG nova.virt.hardware [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 564.720407] env[63531]: DEBUG nova.virt.hardware [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 564.720407] env[63531]: DEBUG nova.virt.hardware [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 564.720630] env[63531]: DEBUG nova.virt.hardware [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 564.720630] env[63531]: DEBUG nova.virt.hardware [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 564.720630] env[63531]: DEBUG nova.virt.hardware [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 564.721372] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7be7d66-3ede-497f-b85a-e0c14cf70681 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.735550] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41ea4f56-34a8-4d92-83fd-688e9b269be8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.750377] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] [instance: 5b86b63a-f6c1-4983-a3cc-3cde7a10d17b] Instance VIF info [] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 564.757806] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Creating folder: Project (351aa55b4d204e3081bd82d1a58c2315). Parent ref: group-v244585. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 564.760279] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6a9f80cd-73ef-4e89-b6b5-6449406c451e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.769021] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Created folder: Project (351aa55b4d204e3081bd82d1a58c2315) in parent group-v244585. [ 564.769021] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Creating folder: Instances. Parent ref: group-v244591. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 564.769021] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b51bd684-584a-4297-ad88-c994841f47e4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.781615] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Created folder: Instances in parent group-v244591. [ 564.781921] env[63531]: DEBUG oslo.service.loopingcall [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 564.782180] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5b86b63a-f6c1-4983-a3cc-3cde7a10d17b] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 564.782438] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-78c84170-2986-42d8-a69e-b567e49a9013 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.801361] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 564.801361] env[63531]: value = "task-1117477" [ 564.801361] env[63531]: _type = "Task" [ 564.801361] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 564.809451] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117477, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 564.966420] env[63531]: DEBUG oslo_concurrency.lockutils [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Acquiring lock "2b5e3059-2772-496b-95c7-a083a37ef155" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.966711] env[63531]: DEBUG oslo_concurrency.lockutils [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Lock "2b5e3059-2772-496b-95c7-a083a37ef155" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.041825] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] Releasing lock "refresh_cache-35e9c404-bad4-4f8d-8c79-fb7250abb363" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 565.042502] env[63531]: DEBUG nova.compute.manager [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 565.042502] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 565.043357] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-27db7c25-9202-463e-801d-d47c16501eeb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.057210] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cacee43-7031-4780-b99f-bd48c51b9534 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.072329] env[63531]: DEBUG nova.network.neutron [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 565.089900] env[63531]: WARNING nova.virt.vmwareapi.vmops [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 35e9c404-bad4-4f8d-8c79-fb7250abb363 could not be found. [ 565.089900] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 565.089900] env[63531]: INFO nova.compute.manager [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] Took 0.05 seconds to destroy the instance on the hypervisor. [ 565.089900] env[63531]: DEBUG oslo.service.loopingcall [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 565.090382] env[63531]: DEBUG nova.compute.manager [-] [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 565.090496] env[63531]: DEBUG nova.network.neutron [-] [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 565.134579] env[63531]: DEBUG nova.network.neutron [-] [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 565.134579] env[63531]: DEBUG nova.network.neutron [req-d43ea6b9-6932-487b-8c39-018d114936ce req-3dfa0e45-34cf-4189-bfa3-ca3ce0a0a11f service nova] [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 565.185312] env[63531]: DEBUG nova.network.neutron [-] [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 565.291723] env[63531]: DEBUG nova.network.neutron [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 565.315768] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117477, 'name': CreateVM_Task, 'duration_secs': 0.278792} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 565.315894] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5b86b63a-f6c1-4983-a3cc-3cde7a10d17b] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 565.317375] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 565.317509] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 565.317820] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 565.318082] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e95a26d9-bb44-4fc8-a501-402ff9a9dde5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.324320] env[63531]: DEBUG oslo_vmware.api [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Waiting for the task: (returnval){ [ 565.324320] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52f61245-ce4f-0d85-8f10-2aaf77ca8b6a" [ 565.324320] env[63531]: _type = "Task" [ 565.324320] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 565.333600] env[63531]: DEBUG oslo_vmware.api [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52f61245-ce4f-0d85-8f10-2aaf77ca8b6a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 565.351769] env[63531]: DEBUG nova.network.neutron [req-d43ea6b9-6932-487b-8c39-018d114936ce req-3dfa0e45-34cf-4189-bfa3-ca3ce0a0a11f service nova] [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 565.576095] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] Acquiring lock "01a1b3e0-4ad9-4350-b59e-499f1b3412a5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.576373] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] Lock "01a1b3e0-4ad9-4350-b59e-499f1b3412a5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.642824] env[63531]: DEBUG nova.network.neutron [-] [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 565.653257] env[63531]: ERROR nova.compute.manager [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 60020572-c02a-4101-9cec-88afc59a70d9, please check neutron logs for more information. [ 565.653257] env[63531]: ERROR nova.compute.manager Traceback (most recent call last): [ 565.653257] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 565.653257] env[63531]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 565.653257] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 565.653257] env[63531]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 565.653257] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 565.653257] env[63531]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 565.653257] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 565.653257] env[63531]: ERROR nova.compute.manager self.force_reraise() [ 565.653257] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 565.653257] env[63531]: ERROR nova.compute.manager raise self.value [ 565.653257] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 565.653257] env[63531]: ERROR nova.compute.manager updated_port = self._update_port( [ 565.653257] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 565.653257] env[63531]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 565.654097] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 565.654097] env[63531]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 565.654097] env[63531]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 60020572-c02a-4101-9cec-88afc59a70d9, please check neutron logs for more information. [ 565.654097] env[63531]: ERROR nova.compute.manager [ 565.654097] env[63531]: Traceback (most recent call last): [ 565.654097] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 565.654097] env[63531]: listener.cb(fileno) [ 565.654097] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 565.654097] env[63531]: result = function(*args, **kwargs) [ 565.654097] env[63531]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 565.654097] env[63531]: return func(*args, **kwargs) [ 565.654097] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 565.654097] env[63531]: raise e [ 565.654097] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 565.654097] env[63531]: nwinfo = self.network_api.allocate_for_instance( [ 565.654097] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 565.654097] env[63531]: created_port_ids = self._update_ports_for_instance( [ 565.654097] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 565.654097] env[63531]: with excutils.save_and_reraise_exception(): [ 565.654097] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 565.654097] env[63531]: self.force_reraise() [ 565.654097] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 565.654097] env[63531]: raise self.value [ 565.654097] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 565.654097] env[63531]: updated_port = self._update_port( [ 565.654097] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 565.654097] env[63531]: _ensure_no_port_binding_failure(port) [ 565.654097] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 565.654097] env[63531]: raise exception.PortBindingFailed(port_id=port['id']) [ 565.655255] env[63531]: nova.exception.PortBindingFailed: Binding failed for port 60020572-c02a-4101-9cec-88afc59a70d9, please check neutron logs for more information. [ 565.655255] env[63531]: Removing descriptor: 17 [ 565.655255] env[63531]: ERROR nova.compute.manager [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 60020572-c02a-4101-9cec-88afc59a70d9, please check neutron logs for more information. [ 565.655255] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] Traceback (most recent call last): [ 565.655255] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 565.655255] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] yield resources [ 565.655255] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 565.655255] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] self.driver.spawn(context, instance, image_meta, [ 565.655255] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 565.655255] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 565.655255] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 565.655255] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] vm_ref = self.build_virtual_machine(instance, [ 565.655636] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 565.655636] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] vif_infos = vmwarevif.get_vif_info(self._session, [ 565.655636] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 565.655636] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] for vif in network_info: [ 565.655636] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 565.655636] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] return self._sync_wrapper(fn, *args, **kwargs) [ 565.655636] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 565.655636] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] self.wait() [ 565.655636] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 565.655636] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] self[:] = self._gt.wait() [ 565.655636] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 565.655636] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] return self._exit_event.wait() [ 565.655636] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 565.656041] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] result = hub.switch() [ 565.656041] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 565.656041] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] return self.greenlet.switch() [ 565.656041] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 565.656041] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] result = function(*args, **kwargs) [ 565.656041] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 565.656041] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] return func(*args, **kwargs) [ 565.656041] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 565.656041] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] raise e [ 565.656041] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 565.656041] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] nwinfo = self.network_api.allocate_for_instance( [ 565.656041] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 565.656041] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] created_port_ids = self._update_ports_for_instance( [ 565.656646] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 565.656646] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] with excutils.save_and_reraise_exception(): [ 565.656646] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 565.656646] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] self.force_reraise() [ 565.656646] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 565.656646] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] raise self.value [ 565.656646] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 565.656646] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] updated_port = self._update_port( [ 565.656646] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 565.656646] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] _ensure_no_port_binding_failure(port) [ 565.656646] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 565.656646] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] raise exception.PortBindingFailed(port_id=port['id']) [ 565.658833] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] nova.exception.PortBindingFailed: Binding failed for port 60020572-c02a-4101-9cec-88afc59a70d9, please check neutron logs for more information. [ 565.658833] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] [ 565.658833] env[63531]: INFO nova.compute.manager [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] Terminating instance [ 565.660341] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] Acquiring lock "refresh_cache-5287b3ed-12ec-47fb-93a2-94038cb2e7b1" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 565.660542] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] Acquired lock "refresh_cache-5287b3ed-12ec-47fb-93a2-94038cb2e7b1" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 565.660768] env[63531]: DEBUG nova.network.neutron [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 565.694546] env[63531]: INFO nova.compute.manager [-] [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] Took 1.07 seconds to deallocate network for instance. [ 565.697129] env[63531]: DEBUG nova.compute.claims [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] Aborting claim: {{(pid=63531) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 565.697325] env[63531]: DEBUG oslo_concurrency.lockutils [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.736473] env[63531]: DEBUG oslo_concurrency.lockutils [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] Acquiring lock "8ac0012a-0855-4b73-a1e4-0a68a1336404" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.736473] env[63531]: DEBUG oslo_concurrency.lockutils [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] Lock "8ac0012a-0855-4b73-a1e4-0a68a1336404" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.795401] env[63531]: DEBUG oslo_concurrency.lockutils [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] Releasing lock "refresh_cache-ca9c41cf-b5e6-400b-9b6c-34a065722969" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 565.795649] env[63531]: DEBUG nova.compute.manager [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 565.795832] env[63531]: DEBUG nova.compute.manager [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 565.796137] env[63531]: DEBUG nova.network.neutron [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 565.811108] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb3a467b-81b3-49db-bf34-4d0411a72274 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.814445] env[63531]: DEBUG nova.network.neutron [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 565.822891] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46ae03e5-f012-496c-9986-1b4ec95b5764 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.860396] env[63531]: DEBUG oslo_vmware.api [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52f61245-ce4f-0d85-8f10-2aaf77ca8b6a, 'name': SearchDatastore_Task, 'duration_secs': 0.009909} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 565.860768] env[63531]: INFO nova.compute.manager [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Rebuilding instance [ 565.862857] env[63531]: DEBUG oslo_concurrency.lockutils [req-d43ea6b9-6932-487b-8c39-018d114936ce req-3dfa0e45-34cf-4189-bfa3-ca3ce0a0a11f service nova] Releasing lock "refresh_cache-6e963227-64f7-414d-88ec-cee5d5770bba" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 565.863684] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 565.863964] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] [instance: 5b86b63a-f6c1-4983-a3cc-3cde7a10d17b] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 565.864143] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 565.864290] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 565.864465] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 565.865283] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a6cdc4f-10c2-4d5d-b913-22d450666511 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.867808] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-18aa5838-3f75-4192-a206-abee6682b16c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.875210] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-923e8650-ae2a-4e61-a18a-f6229f75f88d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.881258] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 565.881427] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 565.882559] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-077231be-29ec-4f11-9e64-053c58b7b131 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.893190] env[63531]: DEBUG nova.compute.provider_tree [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 565.902031] env[63531]: DEBUG oslo_vmware.api [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Waiting for the task: (returnval){ [ 565.902031] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52a96b57-9d7d-2c9d-fed7-7a14e29c79d8" [ 565.902031] env[63531]: _type = "Task" [ 565.902031] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 565.911059] env[63531]: DEBUG oslo_vmware.api [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52a96b57-9d7d-2c9d-fed7-7a14e29c79d8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 565.913852] env[63531]: DEBUG nova.compute.manager [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 565.914622] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f73ee516-470d-49c1-b843-386275ef0af8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.149502] env[63531]: INFO nova.compute.manager [-] [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] Took 1.06 seconds to deallocate network for instance. [ 566.150646] env[63531]: DEBUG nova.compute.claims [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] Aborting claim: {{(pid=63531) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 566.150974] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.190836] env[63531]: DEBUG nova.network.neutron [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 566.320655] env[63531]: DEBUG nova.network.neutron [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 566.370245] env[63531]: DEBUG nova.network.neutron [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 566.399903] env[63531]: DEBUG nova.scheduler.client.report [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 566.415795] env[63531]: DEBUG oslo_vmware.api [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52a96b57-9d7d-2c9d-fed7-7a14e29c79d8, 'name': SearchDatastore_Task, 'duration_secs': 0.008633} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 566.417164] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bcf8cdd5-0d7f-40b8-bf5f-eafe928582f2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.424089] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 566.424427] env[63531]: DEBUG oslo_vmware.api [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Waiting for the task: (returnval){ [ 566.424427] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]527c6943-c2c8-4050-4222-5c4b96dd1941" [ 566.424427] env[63531]: _type = "Task" [ 566.424427] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 566.425230] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f2385da8-166c-4c37-ae3a-ead8cd33c571 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.440948] env[63531]: DEBUG oslo_vmware.api [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]527c6943-c2c8-4050-4222-5c4b96dd1941, 'name': SearchDatastore_Task, 'duration_secs': 0.012383} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 566.444111] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 566.444111] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 5b86b63a-f6c1-4983-a3cc-3cde7a10d17b/5b86b63a-f6c1-4983-a3cc-3cde7a10d17b.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 566.444111] env[63531]: DEBUG oslo_vmware.api [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Waiting for the task: (returnval){ [ 566.444111] env[63531]: value = "task-1117478" [ 566.444111] env[63531]: _type = "Task" [ 566.444111] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 566.444111] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a23845fc-dc01-447e-a8e3-44dbc76f5941 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.454575] env[63531]: DEBUG oslo_vmware.api [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': task-1117478, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 566.455853] env[63531]: DEBUG oslo_vmware.api [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Waiting for the task: (returnval){ [ 566.455853] env[63531]: value = "task-1117479" [ 566.455853] env[63531]: _type = "Task" [ 566.455853] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 566.464138] env[63531]: DEBUG oslo_vmware.api [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Task: {'id': task-1117479, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 566.827820] env[63531]: INFO nova.compute.manager [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] [instance: ca9c41cf-b5e6-400b-9b6c-34a065722969] Took 1.03 seconds to deallocate network for instance. [ 566.849210] env[63531]: DEBUG oslo_concurrency.lockutils [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Acquiring lock "3a02946a-64e9-41b7-95d7-0b32766d735c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.849490] env[63531]: DEBUG oslo_concurrency.lockutils [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Lock "3a02946a-64e9-41b7-95d7-0b32766d735c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 566.874948] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] Releasing lock "refresh_cache-5287b3ed-12ec-47fb-93a2-94038cb2e7b1" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 566.875342] env[63531]: DEBUG nova.compute.manager [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 566.875826] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 566.875826] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-51df7543-7862-4de9-8468-79e8965ef121 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.890056] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aef0c685-457e-4fc7-b48c-6c7b50aedcc5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.913363] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.395s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 566.915271] env[63531]: DEBUG nova.compute.manager [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 566.917426] env[63531]: DEBUG oslo_concurrency.lockutils [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 10.674s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 566.926616] env[63531]: DEBUG nova.compute.manager [req-b7812830-1b5e-41df-a2b6-4114e638a2bb req-9bc655cf-b9d2-4197-9f3b-331a07509453 service nova] [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] Received event network-changed-e006227b-9069-49b7-a244-083d6aed19d7 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 566.926856] env[63531]: DEBUG nova.compute.manager [req-b7812830-1b5e-41df-a2b6-4114e638a2bb req-9bc655cf-b9d2-4197-9f3b-331a07509453 service nova] [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] Refreshing instance network info cache due to event network-changed-e006227b-9069-49b7-a244-083d6aed19d7. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 566.927101] env[63531]: DEBUG oslo_concurrency.lockutils [req-b7812830-1b5e-41df-a2b6-4114e638a2bb req-9bc655cf-b9d2-4197-9f3b-331a07509453 service nova] Acquiring lock "refresh_cache-35e9c404-bad4-4f8d-8c79-fb7250abb363" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 566.928508] env[63531]: DEBUG oslo_concurrency.lockutils [req-b7812830-1b5e-41df-a2b6-4114e638a2bb req-9bc655cf-b9d2-4197-9f3b-331a07509453 service nova] Acquired lock "refresh_cache-35e9c404-bad4-4f8d-8c79-fb7250abb363" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 566.928508] env[63531]: DEBUG nova.network.neutron [req-b7812830-1b5e-41df-a2b6-4114e638a2bb req-9bc655cf-b9d2-4197-9f3b-331a07509453 service nova] [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] Refreshing network info cache for port e006227b-9069-49b7-a244-083d6aed19d7 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 566.941219] env[63531]: WARNING nova.virt.vmwareapi.vmops [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5287b3ed-12ec-47fb-93a2-94038cb2e7b1 could not be found. [ 566.941449] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 566.941707] env[63531]: INFO nova.compute.manager [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] Took 0.07 seconds to destroy the instance on the hypervisor. [ 566.941971] env[63531]: DEBUG oslo.service.loopingcall [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 566.942587] env[63531]: DEBUG nova.compute.manager [-] [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 566.942697] env[63531]: DEBUG nova.network.neutron [-] [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 566.960153] env[63531]: DEBUG oslo_vmware.api [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': task-1117478, 'name': PowerOffVM_Task, 'duration_secs': 0.266219} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 566.964073] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 566.964348] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 566.967118] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88eaf2c3-2180-4af7-b4d4-a8643ecfd05a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.981048] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 566.981048] env[63531]: DEBUG oslo_vmware.api [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Task: {'id': task-1117479, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 566.981048] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4cb6854d-ad41-4179-aca8-582616538844 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.991563] env[63531]: DEBUG nova.network.neutron [-] [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 567.010346] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 567.011161] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 567.011161] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Deleting the datastore file [datastore1] 4410b013-868a-41e7-9dfb-584fb2d9d142 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 567.011161] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b26d2a91-e17e-47b1-8cca-5f53c30c898a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.022705] env[63531]: DEBUG oslo_vmware.api [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Waiting for the task: (returnval){ [ 567.022705] env[63531]: value = "task-1117481" [ 567.022705] env[63531]: _type = "Task" [ 567.022705] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 567.043891] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] Acquiring lock "fe881dbe-6d44-4e31-a537-4aa9dfe44ff4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 567.044030] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] Lock "fe881dbe-6d44-4e31-a537-4aa9dfe44ff4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 567.049914] env[63531]: DEBUG oslo_vmware.api [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': task-1117481, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 567.413092] env[63531]: DEBUG nova.compute.manager [req-dc4d2477-f50d-4d45-a45b-f936ec4fd2b0 req-1b935eba-0517-4b5f-8a60-bc7ba34c2248 service nova] [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] Received event network-vif-deleted-e70dcd41-edfa-4975-a540-6e7d7bbd6280 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 567.413366] env[63531]: DEBUG nova.compute.manager [req-dc4d2477-f50d-4d45-a45b-f936ec4fd2b0 req-1b935eba-0517-4b5f-8a60-bc7ba34c2248 service nova] [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] Received event network-changed-60020572-c02a-4101-9cec-88afc59a70d9 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 567.414374] env[63531]: DEBUG nova.compute.manager [req-dc4d2477-f50d-4d45-a45b-f936ec4fd2b0 req-1b935eba-0517-4b5f-8a60-bc7ba34c2248 service nova] [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] Refreshing instance network info cache due to event network-changed-60020572-c02a-4101-9cec-88afc59a70d9. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 567.414656] env[63531]: DEBUG oslo_concurrency.lockutils [req-dc4d2477-f50d-4d45-a45b-f936ec4fd2b0 req-1b935eba-0517-4b5f-8a60-bc7ba34c2248 service nova] Acquiring lock "refresh_cache-5287b3ed-12ec-47fb-93a2-94038cb2e7b1" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 567.414813] env[63531]: DEBUG oslo_concurrency.lockutils [req-dc4d2477-f50d-4d45-a45b-f936ec4fd2b0 req-1b935eba-0517-4b5f-8a60-bc7ba34c2248 service nova] Acquired lock "refresh_cache-5287b3ed-12ec-47fb-93a2-94038cb2e7b1" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 567.414976] env[63531]: DEBUG nova.network.neutron [req-dc4d2477-f50d-4d45-a45b-f936ec4fd2b0 req-1b935eba-0517-4b5f-8a60-bc7ba34c2248 service nova] [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] Refreshing network info cache for port 60020572-c02a-4101-9cec-88afc59a70d9 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 567.426796] env[63531]: DEBUG nova.compute.utils [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 567.433284] env[63531]: DEBUG nova.compute.manager [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 567.433284] env[63531]: DEBUG nova.network.neutron [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 567.470334] env[63531]: DEBUG oslo_vmware.api [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Task: {'id': task-1117479, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.558396} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 567.471048] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 5b86b63a-f6c1-4983-a3cc-3cde7a10d17b/5b86b63a-f6c1-4983-a3cc-3cde7a10d17b.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 567.471390] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] [instance: 5b86b63a-f6c1-4983-a3cc-3cde7a10d17b] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 567.476041] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-44a48f20-d729-48de-9729-8fa1a2a4b2f5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.481151] env[63531]: DEBUG nova.network.neutron [req-b7812830-1b5e-41df-a2b6-4114e638a2bb req-9bc655cf-b9d2-4197-9f3b-331a07509453 service nova] [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 567.497814] env[63531]: DEBUG nova.network.neutron [-] [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 567.499646] env[63531]: DEBUG oslo_vmware.api [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Waiting for the task: (returnval){ [ 567.499646] env[63531]: value = "task-1117482" [ 567.499646] env[63531]: _type = "Task" [ 567.499646] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 567.510156] env[63531]: DEBUG oslo_concurrency.lockutils [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] Acquiring lock "3c513284-4860-46b2-8ba1-e185efe0db86" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 567.510563] env[63531]: DEBUG oslo_concurrency.lockutils [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] Lock "3c513284-4860-46b2-8ba1-e185efe0db86" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 567.523351] env[63531]: DEBUG oslo_vmware.api [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Task: {'id': task-1117482, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 567.547808] env[63531]: DEBUG oslo_vmware.api [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': task-1117481, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.449546} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 567.548052] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 567.548248] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 567.548449] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 567.603778] env[63531]: DEBUG nova.policy [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2e047a775d4642dca798c5d624ccbc89', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5f0d10f679de4fa4bd9d725951e536c1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 567.695108] env[63531]: DEBUG nova.network.neutron [req-b7812830-1b5e-41df-a2b6-4114e638a2bb req-9bc655cf-b9d2-4197-9f3b-331a07509453 service nova] [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 567.777860] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-235b4aef-cf28-47b6-906d-2059a2b46722 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.787848] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4800bae1-ef9e-48c8-b823-9fffb2d8b24a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.821235] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32123e6e-38a2-42f2-8de2-20f7489ad006 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.829156] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c741495f-778b-4cac-a6c9-4831ec798bca {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.845798] env[63531]: DEBUG nova.compute.provider_tree [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 567.881286] env[63531]: INFO nova.scheduler.client.report [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] Deleted allocations for instance ca9c41cf-b5e6-400b-9b6c-34a065722969 [ 567.934518] env[63531]: DEBUG nova.compute.manager [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 567.951087] env[63531]: DEBUG nova.network.neutron [req-dc4d2477-f50d-4d45-a45b-f936ec4fd2b0 req-1b935eba-0517-4b5f-8a60-bc7ba34c2248 service nova] [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 568.004027] env[63531]: INFO nova.compute.manager [-] [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] Took 1.06 seconds to deallocate network for instance. [ 568.010991] env[63531]: DEBUG nova.compute.claims [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] Aborting claim: {{(pid=63531) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 568.011805] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 568.011805] env[63531]: DEBUG oslo_vmware.api [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Task: {'id': task-1117482, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074982} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 568.011805] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] [instance: 5b86b63a-f6c1-4983-a3cc-3cde7a10d17b] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 568.013762] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47dd6175-68f6-4827-be9a-4a895b837232 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.041255] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] [instance: 5b86b63a-f6c1-4983-a3cc-3cde7a10d17b] Reconfiguring VM instance instance-00000009 to attach disk [datastore1] 5b86b63a-f6c1-4983-a3cc-3cde7a10d17b/5b86b63a-f6c1-4983-a3cc-3cde7a10d17b.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 568.042092] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8de159cf-b7e5-4431-9a28-169b6e0e2ef4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.069891] env[63531]: DEBUG oslo_vmware.api [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Waiting for the task: (returnval){ [ 568.069891] env[63531]: value = "task-1117484" [ 568.069891] env[63531]: _type = "Task" [ 568.069891] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 568.082178] env[63531]: DEBUG oslo_vmware.api [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Task: {'id': task-1117484, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 568.197525] env[63531]: DEBUG oslo_concurrency.lockutils [req-b7812830-1b5e-41df-a2b6-4114e638a2bb req-9bc655cf-b9d2-4197-9f3b-331a07509453 service nova] Releasing lock "refresh_cache-35e9c404-bad4-4f8d-8c79-fb7250abb363" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 568.198781] env[63531]: DEBUG nova.compute.manager [req-b7812830-1b5e-41df-a2b6-4114e638a2bb req-9bc655cf-b9d2-4197-9f3b-331a07509453 service nova] [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] Received event network-vif-deleted-e006227b-9069-49b7-a244-083d6aed19d7 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 568.280044] env[63531]: DEBUG nova.network.neutron [req-dc4d2477-f50d-4d45-a45b-f936ec4fd2b0 req-1b935eba-0517-4b5f-8a60-bc7ba34c2248 service nova] [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 568.351346] env[63531]: DEBUG nova.scheduler.client.report [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 568.395770] env[63531]: DEBUG oslo_concurrency.lockutils [None req-68b36d6d-1a64-4c1d-9ae6-ad22867b2612 tempest-ServersTestJSON-1132998076 tempest-ServersTestJSON-1132998076-project-member] Lock "ca9c41cf-b5e6-400b-9b6c-34a065722969" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.084s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 568.415531] env[63531]: DEBUG nova.network.neutron [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] Successfully created port: 670e808d-7316-4724-9751-f9f2a9c1c9ac {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 568.446198] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Acquiring lock "56872ede-96f0-48ca-b289-86784f37527d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 568.446198] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Lock "56872ede-96f0-48ca-b289-86784f37527d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 568.583154] env[63531]: DEBUG oslo_vmware.api [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Task: {'id': task-1117484, 'name': ReconfigVM_Task, 'duration_secs': 0.335293} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 568.584407] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] [instance: 5b86b63a-f6c1-4983-a3cc-3cde7a10d17b] Reconfigured VM instance instance-00000009 to attach disk [datastore1] 5b86b63a-f6c1-4983-a3cc-3cde7a10d17b/5b86b63a-f6c1-4983-a3cc-3cde7a10d17b.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 568.584704] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7485c531-85ca-4664-8f57-06947a147594 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.592021] env[63531]: DEBUG oslo_vmware.api [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Waiting for the task: (returnval){ [ 568.592021] env[63531]: value = "task-1117485" [ 568.592021] env[63531]: _type = "Task" [ 568.592021] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 568.600042] env[63531]: DEBUG oslo_vmware.api [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Task: {'id': task-1117485, 'name': Rename_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 568.609808] env[63531]: DEBUG nova.virt.hardware [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 568.610420] env[63531]: DEBUG nova.virt.hardware [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 568.610807] env[63531]: DEBUG nova.virt.hardware [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 568.611738] env[63531]: DEBUG nova.virt.hardware [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 568.611738] env[63531]: DEBUG nova.virt.hardware [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 568.611738] env[63531]: DEBUG nova.virt.hardware [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 568.611738] env[63531]: DEBUG nova.virt.hardware [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 568.611876] env[63531]: DEBUG nova.virt.hardware [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 568.612018] env[63531]: DEBUG nova.virt.hardware [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 568.612196] env[63531]: DEBUG nova.virt.hardware [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 568.612372] env[63531]: DEBUG nova.virt.hardware [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 568.613923] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-389f5f86-41dc-4613-af0a-9416d60bcd15 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.625605] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87ed7a99-8431-43a5-b93c-64dc9f78c94c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.642823] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Instance VIF info [] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 568.649357] env[63531]: DEBUG oslo.service.loopingcall [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 568.649626] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 568.649846] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-006c720c-739b-48c5-99e3-9d223660104e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.667576] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 568.667576] env[63531]: value = "task-1117486" [ 568.667576] env[63531]: _type = "Task" [ 568.667576] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 568.675385] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117486, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 568.784285] env[63531]: DEBUG oslo_concurrency.lockutils [req-dc4d2477-f50d-4d45-a45b-f936ec4fd2b0 req-1b935eba-0517-4b5f-8a60-bc7ba34c2248 service nova] Releasing lock "refresh_cache-5287b3ed-12ec-47fb-93a2-94038cb2e7b1" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 568.784285] env[63531]: DEBUG nova.compute.manager [req-dc4d2477-f50d-4d45-a45b-f936ec4fd2b0 req-1b935eba-0517-4b5f-8a60-bc7ba34c2248 service nova] [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] Received event network-vif-deleted-60020572-c02a-4101-9cec-88afc59a70d9 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 568.856674] env[63531]: DEBUG oslo_concurrency.lockutils [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.939s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 568.857488] env[63531]: ERROR nova.compute.manager [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 554f8db0-c7eb-4bbd-b6e3-2aa10b61f54b, please check neutron logs for more information. [ 568.857488] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] Traceback (most recent call last): [ 568.857488] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 568.857488] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] self.driver.spawn(context, instance, image_meta, [ 568.857488] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 568.857488] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 568.857488] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 568.857488] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] vm_ref = self.build_virtual_machine(instance, [ 568.857488] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 568.857488] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] vif_infos = vmwarevif.get_vif_info(self._session, [ 568.857488] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 568.857773] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] for vif in network_info: [ 568.857773] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 568.857773] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] return self._sync_wrapper(fn, *args, **kwargs) [ 568.857773] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 568.857773] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] self.wait() [ 568.857773] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 568.857773] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] self[:] = self._gt.wait() [ 568.857773] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 568.857773] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] return self._exit_event.wait() [ 568.857773] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 568.857773] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] result = hub.switch() [ 568.857773] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 568.857773] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] return self.greenlet.switch() [ 568.858106] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 568.858106] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] result = function(*args, **kwargs) [ 568.858106] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 568.858106] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] return func(*args, **kwargs) [ 568.858106] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 568.858106] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] raise e [ 568.858106] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 568.858106] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] nwinfo = self.network_api.allocate_for_instance( [ 568.858106] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 568.858106] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] created_port_ids = self._update_ports_for_instance( [ 568.858106] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 568.858106] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] with excutils.save_and_reraise_exception(): [ 568.858106] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 568.858450] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] self.force_reraise() [ 568.858450] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 568.858450] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] raise self.value [ 568.858450] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 568.858450] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] updated_port = self._update_port( [ 568.858450] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 568.858450] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] _ensure_no_port_binding_failure(port) [ 568.858450] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 568.858450] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] raise exception.PortBindingFailed(port_id=port['id']) [ 568.858450] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] nova.exception.PortBindingFailed: Binding failed for port 554f8db0-c7eb-4bbd-b6e3-2aa10b61f54b, please check neutron logs for more information. [ 568.858450] env[63531]: ERROR nova.compute.manager [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] [ 568.858723] env[63531]: DEBUG nova.compute.utils [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] Binding failed for port 554f8db0-c7eb-4bbd-b6e3-2aa10b61f54b, please check neutron logs for more information. {{(pid=63531) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 568.860475] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 10.036s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 568.860475] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 568.860475] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63531) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 568.860475] env[63531]: DEBUG oslo_concurrency.lockutils [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 8.559s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 568.863324] env[63531]: DEBUG nova.compute.manager [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] Build of instance baa18517-333c-4ebc-8de6-e2fc9576b9fd was re-scheduled: Binding failed for port 554f8db0-c7eb-4bbd-b6e3-2aa10b61f54b, please check neutron logs for more information. {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 568.863786] env[63531]: DEBUG nova.compute.manager [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] Unplugging VIFs for instance {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 568.864468] env[63531]: DEBUG oslo_concurrency.lockutils [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] Acquiring lock "refresh_cache-baa18517-333c-4ebc-8de6-e2fc9576b9fd" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 568.864468] env[63531]: DEBUG oslo_concurrency.lockutils [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] Acquired lock "refresh_cache-baa18517-333c-4ebc-8de6-e2fc9576b9fd" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 568.864468] env[63531]: DEBUG nova.network.neutron [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 568.866280] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc6286cc-7cd2-4fd1-be6b-9a9676c038e8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.877466] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69b2d7c8-287f-486c-bb79-86942ad69a21 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.900524] env[63531]: DEBUG nova.compute.manager [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 568.908497] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0aad72a-7d66-4e22-9fdb-2f0da8ae41a0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.918467] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd01325b-b9d3-4db0-b6eb-dbd1dce6650f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.965572] env[63531]: DEBUG nova.compute.manager [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 568.968094] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181584MB free_disk=170GB free_vcpus=48 pci_devices=None {{(pid=63531) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 568.968240] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 568.995549] env[63531]: DEBUG nova.virt.hardware [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 568.995803] env[63531]: DEBUG nova.virt.hardware [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 568.995990] env[63531]: DEBUG nova.virt.hardware [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 568.996253] env[63531]: DEBUG nova.virt.hardware [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 568.996439] env[63531]: DEBUG nova.virt.hardware [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 568.996593] env[63531]: DEBUG nova.virt.hardware [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 568.998504] env[63531]: DEBUG nova.virt.hardware [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 568.998504] env[63531]: DEBUG nova.virt.hardware [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 568.998504] env[63531]: DEBUG nova.virt.hardware [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 568.998504] env[63531]: DEBUG nova.virt.hardware [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 568.998504] env[63531]: DEBUG nova.virt.hardware [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 568.998714] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad909345-7ecd-4686-adec-09dfac399ac7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.008346] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28c6bfc7-4eb7-42c6-85c5-494edf3efd59 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.102702] env[63531]: DEBUG oslo_vmware.api [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Task: {'id': task-1117485, 'name': Rename_Task, 'duration_secs': 0.22048} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 569.103140] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] [instance: 5b86b63a-f6c1-4983-a3cc-3cde7a10d17b] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 569.103481] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3652d14e-27fb-4b24-bcfe-3007a6f5055d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.110751] env[63531]: DEBUG oslo_vmware.api [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Waiting for the task: (returnval){ [ 569.110751] env[63531]: value = "task-1117487" [ 569.110751] env[63531]: _type = "Task" [ 569.110751] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 569.125337] env[63531]: DEBUG oslo_vmware.api [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Task: {'id': task-1117487, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 569.177818] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117486, 'name': CreateVM_Task, 'duration_secs': 0.30494} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 569.177991] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 569.178857] env[63531]: DEBUG oslo_vmware.service [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5531c99-297a-499f-86d4-847a9eb7fddb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.189149] env[63531]: DEBUG oslo_concurrency.lockutils [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 569.189149] env[63531]: DEBUG oslo_concurrency.lockutils [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 569.191013] env[63531]: DEBUG oslo_concurrency.lockutils [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 569.191013] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a1807072-45ce-4506-9576-2dc10e3b9c75 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.196900] env[63531]: DEBUG oslo_vmware.api [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Waiting for the task: (returnval){ [ 569.196900] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]527ad838-2be6-d7ff-53de-62826899fb9c" [ 569.196900] env[63531]: _type = "Task" [ 569.196900] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 569.206472] env[63531]: DEBUG oslo_vmware.api [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]527ad838-2be6-d7ff-53de-62826899fb9c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 569.403247] env[63531]: DEBUG nova.network.neutron [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 569.431795] env[63531]: DEBUG oslo_concurrency.lockutils [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 569.511547] env[63531]: DEBUG nova.network.neutron [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 569.622198] env[63531]: DEBUG oslo_vmware.api [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Task: {'id': task-1117487, 'name': PowerOnVM_Task, 'duration_secs': 0.479757} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 569.622502] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] [instance: 5b86b63a-f6c1-4983-a3cc-3cde7a10d17b] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 569.623064] env[63531]: INFO nova.compute.manager [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] [instance: 5b86b63a-f6c1-4983-a3cc-3cde7a10d17b] Took 4.94 seconds to spawn the instance on the hypervisor. [ 569.623064] env[63531]: DEBUG nova.compute.manager [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] [instance: 5b86b63a-f6c1-4983-a3cc-3cde7a10d17b] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 569.623759] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cca84f77-99d7-4f61-9688-2583f480c2ef {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.707759] env[63531]: DEBUG oslo_concurrency.lockutils [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 569.708043] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 569.708336] env[63531]: DEBUG oslo_concurrency.lockutils [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 569.708461] env[63531]: DEBUG oslo_concurrency.lockutils [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 569.708621] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 569.711288] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bb476f39-7c92-4645-b061-6a098035d5ee {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.724807] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 569.724807] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 569.724807] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fb7c2c3-ef27-4b90-8f5d-73c20150b0ed {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.732188] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e4e6ad24-dbc0-4303-9d08-2958b2e0d7cd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.741570] env[63531]: DEBUG oslo_vmware.api [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Waiting for the task: (returnval){ [ 569.741570] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]528b34ec-c632-ecc8-1dcd-50bc5a50d8c1" [ 569.741570] env[63531]: _type = "Task" [ 569.741570] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 569.751740] env[63531]: DEBUG oslo_vmware.api [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]528b34ec-c632-ecc8-1dcd-50bc5a50d8c1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 569.813547] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-478d52f1-5460-4fb9-8142-74732ce8022b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.821872] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a83e351-d358-4888-95ba-5053fab08942 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.855153] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83f0b96f-bfff-4ba7-97f7-fc1b9f3b853e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.863906] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ae0de48-0e4c-450a-bd18-bd0972248127 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.879996] env[63531]: DEBUG nova.compute.provider_tree [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 569.979890] env[63531]: ERROR nova.compute.manager [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 670e808d-7316-4724-9751-f9f2a9c1c9ac, please check neutron logs for more information. [ 569.979890] env[63531]: ERROR nova.compute.manager Traceback (most recent call last): [ 569.979890] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 569.979890] env[63531]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 569.979890] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 569.979890] env[63531]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 569.979890] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 569.979890] env[63531]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 569.979890] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 569.979890] env[63531]: ERROR nova.compute.manager self.force_reraise() [ 569.979890] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 569.979890] env[63531]: ERROR nova.compute.manager raise self.value [ 569.979890] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 569.979890] env[63531]: ERROR nova.compute.manager updated_port = self._update_port( [ 569.979890] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 569.979890] env[63531]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 569.980419] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 569.980419] env[63531]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 569.980419] env[63531]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 670e808d-7316-4724-9751-f9f2a9c1c9ac, please check neutron logs for more information. [ 569.980419] env[63531]: ERROR nova.compute.manager [ 569.980419] env[63531]: Traceback (most recent call last): [ 569.980419] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 569.980419] env[63531]: listener.cb(fileno) [ 569.980419] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 569.980419] env[63531]: result = function(*args, **kwargs) [ 569.980419] env[63531]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 569.980419] env[63531]: return func(*args, **kwargs) [ 569.980419] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 569.980419] env[63531]: raise e [ 569.980419] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 569.980419] env[63531]: nwinfo = self.network_api.allocate_for_instance( [ 569.980419] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 569.980419] env[63531]: created_port_ids = self._update_ports_for_instance( [ 569.980419] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 569.980419] env[63531]: with excutils.save_and_reraise_exception(): [ 569.980419] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 569.980419] env[63531]: self.force_reraise() [ 569.980419] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 569.980419] env[63531]: raise self.value [ 569.980419] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 569.980419] env[63531]: updated_port = self._update_port( [ 569.980419] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 569.980419] env[63531]: _ensure_no_port_binding_failure(port) [ 569.980419] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 569.980419] env[63531]: raise exception.PortBindingFailed(port_id=port['id']) [ 569.981990] env[63531]: nova.exception.PortBindingFailed: Binding failed for port 670e808d-7316-4724-9751-f9f2a9c1c9ac, please check neutron logs for more information. [ 569.981990] env[63531]: Removing descriptor: 17 [ 569.981990] env[63531]: ERROR nova.compute.manager [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 670e808d-7316-4724-9751-f9f2a9c1c9ac, please check neutron logs for more information. [ 569.981990] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] Traceback (most recent call last): [ 569.981990] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 569.981990] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] yield resources [ 569.981990] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 569.981990] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] self.driver.spawn(context, instance, image_meta, [ 569.981990] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 569.981990] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] self._vmops.spawn(context, instance, image_meta, injected_files, [ 569.981990] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 569.981990] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] vm_ref = self.build_virtual_machine(instance, [ 569.982316] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 569.982316] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] vif_infos = vmwarevif.get_vif_info(self._session, [ 569.982316] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 569.982316] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] for vif in network_info: [ 569.982316] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 569.982316] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] return self._sync_wrapper(fn, *args, **kwargs) [ 569.982316] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 569.982316] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] self.wait() [ 569.982316] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 569.982316] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] self[:] = self._gt.wait() [ 569.982316] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 569.982316] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] return self._exit_event.wait() [ 569.982316] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 569.982635] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] result = hub.switch() [ 569.982635] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 569.982635] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] return self.greenlet.switch() [ 569.982635] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 569.982635] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] result = function(*args, **kwargs) [ 569.982635] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 569.982635] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] return func(*args, **kwargs) [ 569.982635] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 569.982635] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] raise e [ 569.982635] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 569.982635] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] nwinfo = self.network_api.allocate_for_instance( [ 569.982635] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 569.982635] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] created_port_ids = self._update_ports_for_instance( [ 569.982956] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 569.982956] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] with excutils.save_and_reraise_exception(): [ 569.982956] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 569.982956] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] self.force_reraise() [ 569.982956] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 569.982956] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] raise self.value [ 569.982956] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 569.982956] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] updated_port = self._update_port( [ 569.982956] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 569.982956] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] _ensure_no_port_binding_failure(port) [ 569.982956] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 569.982956] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] raise exception.PortBindingFailed(port_id=port['id']) [ 569.984061] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] nova.exception.PortBindingFailed: Binding failed for port 670e808d-7316-4724-9751-f9f2a9c1c9ac, please check neutron logs for more information. [ 569.984061] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] [ 569.984061] env[63531]: INFO nova.compute.manager [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] Terminating instance [ 569.984061] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Acquiring lock "refresh_cache-6a380967-bca6-47d2-b052-988e6c1004ee" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 569.984061] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Acquired lock "refresh_cache-6a380967-bca6-47d2-b052-988e6c1004ee" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 569.984061] env[63531]: DEBUG nova.network.neutron [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 570.015136] env[63531]: DEBUG oslo_concurrency.lockutils [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] Releasing lock "refresh_cache-baa18517-333c-4ebc-8de6-e2fc9576b9fd" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 570.015383] env[63531]: DEBUG nova.compute.manager [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 570.015541] env[63531]: DEBUG nova.compute.manager [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 570.016183] env[63531]: DEBUG nova.network.neutron [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 570.034684] env[63531]: DEBUG nova.network.neutron [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 570.151613] env[63531]: INFO nova.compute.manager [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] [instance: 5b86b63a-f6c1-4983-a3cc-3cde7a10d17b] Took 19.16 seconds to build instance. [ 570.252954] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Preparing fetch location {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 570.253281] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Creating directory with path [datastore2] vmware_temp/c02c825f-915a-498f-80e4-778f9f67272b/9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 570.253566] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-119bac11-2027-41b6-a978-23884ff769e9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.264981] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Created directory with path [datastore2] vmware_temp/c02c825f-915a-498f-80e4-778f9f67272b/9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 570.265229] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Fetch image to [datastore2] vmware_temp/c02c825f-915a-498f-80e4-778f9f67272b/9d550399-588c-4bd5-8436-eba9b523bd13/tmp-sparse.vmdk {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 570.265390] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Downloading image file data 9d550399-588c-4bd5-8436-eba9b523bd13 to [datastore2] vmware_temp/c02c825f-915a-498f-80e4-778f9f67272b/9d550399-588c-4bd5-8436-eba9b523bd13/tmp-sparse.vmdk on the data store datastore2 {{(pid=63531) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 570.266216] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1eef3e85-bab2-445b-a535-12e0236c3f00 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.275392] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9705bbd-2505-48d0-98b0-528dd00447d0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.290670] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a80467bf-b1e7-4d06-bce9-3f215db20b02 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.320853] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b812ffb4-79d4-48b0-b3b6-76ea3e2da538 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.327297] env[63531]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-44187ca9-b94a-4f5b-9d65-3cc1bb621da2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.349397] env[63531]: DEBUG nova.virt.vmwareapi.images [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Downloading image file data 9d550399-588c-4bd5-8436-eba9b523bd13 to the data store datastore2 {{(pid=63531) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 570.385483] env[63531]: DEBUG nova.scheduler.client.report [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 570.411958] env[63531]: DEBUG nova.compute.manager [req-e3ec5bac-f57b-4573-ab21-f26ae3bd6368 req-ae8849d7-6b22-4d3f-8362-30046aa140dd service nova] [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] Received event network-changed-670e808d-7316-4724-9751-f9f2a9c1c9ac {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 570.412081] env[63531]: DEBUG nova.compute.manager [req-e3ec5bac-f57b-4573-ab21-f26ae3bd6368 req-ae8849d7-6b22-4d3f-8362-30046aa140dd service nova] [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] Refreshing instance network info cache due to event network-changed-670e808d-7316-4724-9751-f9f2a9c1c9ac. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 570.412343] env[63531]: DEBUG oslo_concurrency.lockutils [req-e3ec5bac-f57b-4573-ab21-f26ae3bd6368 req-ae8849d7-6b22-4d3f-8362-30046aa140dd service nova] Acquiring lock "refresh_cache-6a380967-bca6-47d2-b052-988e6c1004ee" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 570.421850] env[63531]: DEBUG oslo_vmware.rw_handles [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/c02c825f-915a-498f-80e4-778f9f67272b/9d550399-588c-4bd5-8436-eba9b523bd13/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=63531) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 570.507694] env[63531]: DEBUG nova.network.neutron [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 570.538649] env[63531]: DEBUG nova.network.neutron [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 570.585241] env[63531]: DEBUG nova.network.neutron [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 570.653120] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1a5141e2-1c74-4bce-ae98-7ec23023a710 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Lock "5b86b63a-f6c1-4983-a3cc-3cde7a10d17b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.669s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 570.894905] env[63531]: DEBUG oslo_concurrency.lockutils [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.034s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 570.895425] env[63531]: ERROR nova.compute.manager [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b832b20c-e3b8-40a9-a4cf-91f8c961fd94, please check neutron logs for more information. [ 570.895425] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] Traceback (most recent call last): [ 570.895425] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 570.895425] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] self.driver.spawn(context, instance, image_meta, [ 570.895425] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 570.895425] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 570.895425] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 570.895425] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] vm_ref = self.build_virtual_machine(instance, [ 570.895425] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 570.895425] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] vif_infos = vmwarevif.get_vif_info(self._session, [ 570.895425] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 570.896152] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] for vif in network_info: [ 570.896152] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 570.896152] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] return self._sync_wrapper(fn, *args, **kwargs) [ 570.896152] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 570.896152] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] self.wait() [ 570.896152] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 570.896152] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] self[:] = self._gt.wait() [ 570.896152] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 570.896152] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] return self._exit_event.wait() [ 570.896152] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 570.896152] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] result = hub.switch() [ 570.896152] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 570.896152] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] return self.greenlet.switch() [ 570.896860] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 570.896860] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] result = function(*args, **kwargs) [ 570.896860] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 570.896860] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] return func(*args, **kwargs) [ 570.896860] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 570.896860] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] raise e [ 570.896860] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 570.896860] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] nwinfo = self.network_api.allocate_for_instance( [ 570.896860] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 570.896860] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] created_port_ids = self._update_ports_for_instance( [ 570.896860] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 570.896860] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] with excutils.save_and_reraise_exception(): [ 570.896860] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 570.897507] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] self.force_reraise() [ 570.897507] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 570.897507] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] raise self.value [ 570.897507] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 570.897507] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] updated_port = self._update_port( [ 570.897507] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 570.897507] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] _ensure_no_port_binding_failure(port) [ 570.897507] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 570.897507] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] raise exception.PortBindingFailed(port_id=port['id']) [ 570.897507] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] nova.exception.PortBindingFailed: Binding failed for port b832b20c-e3b8-40a9-a4cf-91f8c961fd94, please check neutron logs for more information. [ 570.897507] env[63531]: ERROR nova.compute.manager [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] [ 570.898164] env[63531]: DEBUG nova.compute.utils [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] Binding failed for port b832b20c-e3b8-40a9-a4cf-91f8c961fd94, please check neutron logs for more information. {{(pid=63531) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 570.898164] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 6.978s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 570.901598] env[63531]: DEBUG nova.compute.manager [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] Build of instance c552e749-33ec-4093-be8c-ed3784bdc2b5 was re-scheduled: Binding failed for port b832b20c-e3b8-40a9-a4cf-91f8c961fd94, please check neutron logs for more information. {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 570.902066] env[63531]: DEBUG nova.compute.manager [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] Unplugging VIFs for instance {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 570.902370] env[63531]: DEBUG oslo_concurrency.lockutils [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] Acquiring lock "refresh_cache-c552e749-33ec-4093-be8c-ed3784bdc2b5" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 570.902448] env[63531]: DEBUG oslo_concurrency.lockutils [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] Acquired lock "refresh_cache-c552e749-33ec-4093-be8c-ed3784bdc2b5" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 570.902644] env[63531]: DEBUG nova.network.neutron [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 571.041222] env[63531]: INFO nova.compute.manager [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] [instance: baa18517-333c-4ebc-8de6-e2fc9576b9fd] Took 1.03 seconds to deallocate network for instance. [ 571.089205] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Releasing lock "refresh_cache-6a380967-bca6-47d2-b052-988e6c1004ee" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 571.089785] env[63531]: DEBUG nova.compute.manager [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 571.089997] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 571.090360] env[63531]: DEBUG oslo_concurrency.lockutils [req-e3ec5bac-f57b-4573-ab21-f26ae3bd6368 req-ae8849d7-6b22-4d3f-8362-30046aa140dd service nova] Acquired lock "refresh_cache-6a380967-bca6-47d2-b052-988e6c1004ee" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 571.090537] env[63531]: DEBUG nova.network.neutron [req-e3ec5bac-f57b-4573-ab21-f26ae3bd6368 req-ae8849d7-6b22-4d3f-8362-30046aa140dd service nova] [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] Refreshing network info cache for port 670e808d-7316-4724-9751-f9f2a9c1c9ac {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 571.095155] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e8d8d9a4-0c72-442b-b519-d87a4888682a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.112228] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee79ee58-8f16-489e-8b3b-7701013f197d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.141048] env[63531]: WARNING nova.virt.vmwareapi.vmops [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6a380967-bca6-47d2-b052-988e6c1004ee could not be found. [ 571.141301] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 571.141490] env[63531]: INFO nova.compute.manager [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] Took 0.05 seconds to destroy the instance on the hypervisor. [ 571.141788] env[63531]: DEBUG oslo.service.loopingcall [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 571.142588] env[63531]: DEBUG nova.compute.manager [-] [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 571.142686] env[63531]: DEBUG nova.network.neutron [-] [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 571.157952] env[63531]: DEBUG nova.compute.manager [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] [instance: d941a593-7047-436a-b95d-bcb4850829f5] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 571.164424] env[63531]: DEBUG nova.network.neutron [-] [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 571.172122] env[63531]: DEBUG nova.network.neutron [-] [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 571.284192] env[63531]: DEBUG oslo_vmware.rw_handles [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Completed reading data from the image iterator. {{(pid=63531) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 571.284925] env[63531]: DEBUG oslo_vmware.rw_handles [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Closing write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/c02c825f-915a-498f-80e4-778f9f67272b/9d550399-588c-4bd5-8436-eba9b523bd13/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=63531) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 571.357893] env[63531]: DEBUG nova.virt.vmwareapi.images [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Downloaded image file data 9d550399-588c-4bd5-8436-eba9b523bd13 to vmware_temp/c02c825f-915a-498f-80e4-778f9f67272b/9d550399-588c-4bd5-8436-eba9b523bd13/tmp-sparse.vmdk on the data store datastore2 {{(pid=63531) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 571.359720] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Caching image {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 571.360113] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Copying Virtual Disk [datastore2] vmware_temp/c02c825f-915a-498f-80e4-778f9f67272b/9d550399-588c-4bd5-8436-eba9b523bd13/tmp-sparse.vmdk to [datastore2] vmware_temp/c02c825f-915a-498f-80e4-778f9f67272b/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 571.360456] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8f8024e3-12ff-4819-b148-dee1347cc195 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.371348] env[63531]: DEBUG oslo_vmware.api [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Waiting for the task: (returnval){ [ 571.371348] env[63531]: value = "task-1117489" [ 571.371348] env[63531]: _type = "Task" [ 571.371348] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 571.382015] env[63531]: DEBUG oslo_vmware.api [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': task-1117489, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 571.432155] env[63531]: DEBUG nova.network.neutron [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 571.579259] env[63531]: DEBUG nova.network.neutron [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 571.676092] env[63531]: DEBUG nova.network.neutron [req-e3ec5bac-f57b-4573-ab21-f26ae3bd6368 req-ae8849d7-6b22-4d3f-8362-30046aa140dd service nova] [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 571.679133] env[63531]: INFO nova.compute.manager [-] [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] Took 0.54 seconds to deallocate network for instance. [ 571.687651] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 571.689807] env[63531]: DEBUG nova.compute.claims [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] Aborting claim: {{(pid=63531) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 571.690183] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 571.765934] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63d5aa9a-6c24-4f0e-bebe-dd81f964a375 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.775855] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7df3b105-10cf-4589-9c24-11bacef62832 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.812195] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56dd7bf6-c990-4bb2-b929-8225f3282fbf {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.820378] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31b29d24-4d59-4c70-a18f-6f6238880a6a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.825792] env[63531]: DEBUG nova.network.neutron [req-e3ec5bac-f57b-4573-ab21-f26ae3bd6368 req-ae8849d7-6b22-4d3f-8362-30046aa140dd service nova] [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 571.843675] env[63531]: DEBUG nova.compute.provider_tree [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 571.880934] env[63531]: DEBUG oslo_vmware.api [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': task-1117489, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 572.085688] env[63531]: DEBUG oslo_concurrency.lockutils [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] Releasing lock "refresh_cache-c552e749-33ec-4093-be8c-ed3784bdc2b5" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 572.085688] env[63531]: DEBUG nova.compute.manager [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 572.085688] env[63531]: DEBUG nova.compute.manager [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 572.085688] env[63531]: DEBUG nova.network.neutron [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 572.098109] env[63531]: INFO nova.scheduler.client.report [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] Deleted allocations for instance baa18517-333c-4ebc-8de6-e2fc9576b9fd [ 572.119645] env[63531]: DEBUG nova.network.neutron [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 572.328503] env[63531]: DEBUG nova.compute.manager [None req-5b2a3183-dcec-47d6-96f7-1e07ec509ffc tempest-ServerDiagnosticsV248Test-1489378919 tempest-ServerDiagnosticsV248Test-1489378919-project-admin] [instance: 5b86b63a-f6c1-4983-a3cc-3cde7a10d17b] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 572.329344] env[63531]: DEBUG oslo_concurrency.lockutils [req-e3ec5bac-f57b-4573-ab21-f26ae3bd6368 req-ae8849d7-6b22-4d3f-8362-30046aa140dd service nova] Releasing lock "refresh_cache-6a380967-bca6-47d2-b052-988e6c1004ee" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 572.332641] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05f4e554-65cf-49b5-9623-4a75aa268a89 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.340953] env[63531]: INFO nova.compute.manager [None req-5b2a3183-dcec-47d6-96f7-1e07ec509ffc tempest-ServerDiagnosticsV248Test-1489378919 tempest-ServerDiagnosticsV248Test-1489378919-project-admin] [instance: 5b86b63a-f6c1-4983-a3cc-3cde7a10d17b] Retrieving diagnostics [ 572.342349] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dda42a43-3659-4990-b6cd-d67bba8ef503 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.349011] env[63531]: DEBUG nova.scheduler.client.report [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 572.387736] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.490s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 572.388390] env[63531]: ERROR nova.compute.manager [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 73d60e7d-c4d2-475d-aeed-b12ea040e3df, please check neutron logs for more information. [ 572.388390] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] Traceback (most recent call last): [ 572.388390] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 572.388390] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] self.driver.spawn(context, instance, image_meta, [ 572.388390] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 572.388390] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] self._vmops.spawn(context, instance, image_meta, injected_files, [ 572.388390] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 572.388390] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] vm_ref = self.build_virtual_machine(instance, [ 572.388390] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 572.388390] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] vif_infos = vmwarevif.get_vif_info(self._session, [ 572.388390] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 572.393203] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] for vif in network_info: [ 572.393203] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 572.393203] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] return self._sync_wrapper(fn, *args, **kwargs) [ 572.393203] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 572.393203] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] self.wait() [ 572.393203] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 572.393203] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] self[:] = self._gt.wait() [ 572.393203] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 572.393203] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] return self._exit_event.wait() [ 572.393203] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 572.393203] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] result = hub.switch() [ 572.393203] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 572.393203] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] return self.greenlet.switch() [ 572.393606] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 572.393606] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] result = function(*args, **kwargs) [ 572.393606] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 572.393606] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] return func(*args, **kwargs) [ 572.393606] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 572.393606] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] raise e [ 572.393606] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 572.393606] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] nwinfo = self.network_api.allocate_for_instance( [ 572.393606] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 572.393606] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] created_port_ids = self._update_ports_for_instance( [ 572.393606] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 572.393606] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] with excutils.save_and_reraise_exception(): [ 572.393606] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 572.393917] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] self.force_reraise() [ 572.393917] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 572.393917] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] raise self.value [ 572.393917] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 572.393917] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] updated_port = self._update_port( [ 572.393917] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 572.393917] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] _ensure_no_port_binding_failure(port) [ 572.393917] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 572.393917] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] raise exception.PortBindingFailed(port_id=port['id']) [ 572.393917] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] nova.exception.PortBindingFailed: Binding failed for port 73d60e7d-c4d2-475d-aeed-b12ea040e3df, please check neutron logs for more information. [ 572.393917] env[63531]: ERROR nova.compute.manager [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] [ 572.395300] env[63531]: DEBUG nova.compute.utils [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] Binding failed for port 73d60e7d-c4d2-475d-aeed-b12ea040e3df, please check neutron logs for more information. {{(pid=63531) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 572.395300] env[63531]: DEBUG oslo_concurrency.lockutils [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.387s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 572.396234] env[63531]: INFO nova.compute.claims [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 572.402243] env[63531]: DEBUG nova.compute.manager [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] Build of instance d1636eb1-b862-47c9-8729-7a9d475ace01 was re-scheduled: Binding failed for port 73d60e7d-c4d2-475d-aeed-b12ea040e3df, please check neutron logs for more information. {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 572.402243] env[63531]: DEBUG nova.compute.manager [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] Unplugging VIFs for instance {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 572.402243] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Acquiring lock "refresh_cache-d1636eb1-b862-47c9-8729-7a9d475ace01" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 572.402243] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Acquired lock "refresh_cache-d1636eb1-b862-47c9-8729-7a9d475ace01" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 572.402532] env[63531]: DEBUG nova.network.neutron [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 572.407504] env[63531]: DEBUG oslo_vmware.api [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': task-1117489, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.710331} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 572.407948] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Copied Virtual Disk [datastore2] vmware_temp/c02c825f-915a-498f-80e4-778f9f67272b/9d550399-588c-4bd5-8436-eba9b523bd13/tmp-sparse.vmdk to [datastore2] vmware_temp/c02c825f-915a-498f-80e4-778f9f67272b/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 572.408115] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Deleting the datastore file [datastore2] vmware_temp/c02c825f-915a-498f-80e4-778f9f67272b/9d550399-588c-4bd5-8436-eba9b523bd13/tmp-sparse.vmdk {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 572.408356] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c73521d8-59a4-4564-9c32-2a4621027c7a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.416466] env[63531]: DEBUG oslo_vmware.api [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Waiting for the task: (returnval){ [ 572.416466] env[63531]: value = "task-1117491" [ 572.416466] env[63531]: _type = "Task" [ 572.416466] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 572.428514] env[63531]: DEBUG oslo_vmware.api [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': task-1117491, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 572.613773] env[63531]: DEBUG oslo_concurrency.lockutils [None req-d42078dd-3a9e-431a-9039-8a3bfe956c5a tempest-TenantUsagesTestJSON-1950854534 tempest-TenantUsagesTestJSON-1950854534-project-member] Lock "baa18517-333c-4ebc-8de6-e2fc9576b9fd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.622s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 572.621588] env[63531]: DEBUG nova.network.neutron [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 572.929985] env[63531]: DEBUG oslo_vmware.api [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': task-1117491, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.022295} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 572.930891] env[63531]: DEBUG nova.network.neutron [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 572.932960] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 572.933362] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Moving file from [datastore2] vmware_temp/c02c825f-915a-498f-80e4-778f9f67272b/9d550399-588c-4bd5-8436-eba9b523bd13 to [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13. {{(pid=63531) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 572.934122] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-f804264f-3a82-4cc3-ae21-e65121c291cb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.941295] env[63531]: DEBUG oslo_vmware.api [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Waiting for the task: (returnval){ [ 572.941295] env[63531]: value = "task-1117492" [ 572.941295] env[63531]: _type = "Task" [ 572.941295] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 572.950738] env[63531]: DEBUG oslo_vmware.api [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': task-1117492, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 573.044876] env[63531]: DEBUG nova.network.neutron [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 573.117437] env[63531]: DEBUG nova.compute.manager [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 573.128813] env[63531]: INFO nova.compute.manager [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] [instance: c552e749-33ec-4093-be8c-ed3784bdc2b5] Took 1.04 seconds to deallocate network for instance. [ 573.453969] env[63531]: DEBUG oslo_vmware.api [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': task-1117492, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.028999} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 573.455190] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] File moved {{(pid=63531) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 573.455484] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Cleaning up location [datastore2] vmware_temp/c02c825f-915a-498f-80e4-778f9f67272b {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 573.455711] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Deleting the datastore file [datastore2] vmware_temp/c02c825f-915a-498f-80e4-778f9f67272b {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 573.456032] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-79f824e3-85ec-4b72-be52-daa365fa34fb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.465617] env[63531]: DEBUG oslo_vmware.api [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Waiting for the task: (returnval){ [ 573.465617] env[63531]: value = "task-1117493" [ 573.465617] env[63531]: _type = "Task" [ 573.465617] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 573.478287] env[63531]: DEBUG oslo_vmware.api [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': task-1117493, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 573.548342] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Releasing lock "refresh_cache-d1636eb1-b862-47c9-8729-7a9d475ace01" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 573.548583] env[63531]: DEBUG nova.compute.manager [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 573.548747] env[63531]: DEBUG nova.compute.manager [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 573.548915] env[63531]: DEBUG nova.network.neutron [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 573.592594] env[63531]: DEBUG nova.network.neutron [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 573.648657] env[63531]: DEBUG oslo_concurrency.lockutils [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 573.724920] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e008627-286f-461f-99bc-82deefae22e4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.733772] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66d6a0c3-fdf0-4e3e-be34-edac0a16cfda {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.771642] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-432dcb5d-ad00-4124-8573-82cc08c8cbc2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.779338] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e184dbd-ab88-4449-bdca-384b6014f476 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.792959] env[63531]: DEBUG nova.compute.provider_tree [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 573.979474] env[63531]: DEBUG oslo_vmware.api [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': task-1117493, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.025405} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 573.979474] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 573.979720] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3272278b-285a-4da7-a75e-bf15777292e3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.985341] env[63531]: DEBUG oslo_vmware.api [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Waiting for the task: (returnval){ [ 573.985341] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52fb3dfb-f378-e099-b8ee-8f98b3abc58f" [ 573.985341] env[63531]: _type = "Task" [ 573.985341] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 573.994192] env[63531]: DEBUG oslo_vmware.api [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52fb3dfb-f378-e099-b8ee-8f98b3abc58f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 574.094977] env[63531]: DEBUG nova.network.neutron [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 574.182432] env[63531]: INFO nova.scheduler.client.report [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] Deleted allocations for instance c552e749-33ec-4093-be8c-ed3784bdc2b5 [ 574.297450] env[63531]: DEBUG nova.scheduler.client.report [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 574.398659] env[63531]: DEBUG nova.compute.manager [req-84280c5c-0868-43a0-bbf4-46014cbf73d7 req-5b9df555-3348-4939-ba27-decfbaae2df6 service nova] [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] Received event network-vif-deleted-670e808d-7316-4724-9751-f9f2a9c1c9ac {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 574.496258] env[63531]: DEBUG oslo_vmware.api [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52fb3dfb-f378-e099-b8ee-8f98b3abc58f, 'name': SearchDatastore_Task, 'duration_secs': 0.009934} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 574.496258] env[63531]: DEBUG oslo_concurrency.lockutils [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 574.496561] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] 4410b013-868a-41e7-9dfb-584fb2d9d142/4410b013-868a-41e7-9dfb-584fb2d9d142.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 574.497028] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-54c5c8a1-3093-4e23-b06f-7cbf66644597 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.506668] env[63531]: DEBUG oslo_vmware.api [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Waiting for the task: (returnval){ [ 574.506668] env[63531]: value = "task-1117495" [ 574.506668] env[63531]: _type = "Task" [ 574.506668] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 574.515263] env[63531]: DEBUG oslo_vmware.api [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': task-1117495, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 574.601997] env[63531]: INFO nova.compute.manager [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: d1636eb1-b862-47c9-8729-7a9d475ace01] Took 1.05 seconds to deallocate network for instance. [ 574.695899] env[63531]: DEBUG oslo_concurrency.lockutils [None req-69afa30c-bbf5-4c2e-8a70-a3ff6a7f7280 tempest-ServersAdminNegativeTestJSON-945776012 tempest-ServersAdminNegativeTestJSON-945776012-project-member] Lock "c552e749-33ec-4093-be8c-ed3784bdc2b5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.435s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 574.807649] env[63531]: DEBUG oslo_concurrency.lockutils [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.413s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 574.811705] env[63531]: DEBUG nova.compute.manager [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 574.815162] env[63531]: DEBUG oslo_concurrency.lockutils [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 9.117s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 575.018477] env[63531]: DEBUG oslo_vmware.api [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': task-1117495, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 575.200676] env[63531]: DEBUG nova.compute.manager [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 575.321659] env[63531]: DEBUG nova.compute.utils [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 575.323128] env[63531]: DEBUG nova.compute.manager [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 575.323315] env[63531]: DEBUG nova.network.neutron [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 575.414396] env[63531]: DEBUG nova.policy [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0d6482def4c14abca635778264037ba8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9ac21a00cdd845c5b70697bb3326c274', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 575.520579] env[63531]: DEBUG oslo_vmware.api [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': task-1117495, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.55515} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 575.523370] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] 4410b013-868a-41e7-9dfb-584fb2d9d142/4410b013-868a-41e7-9dfb-584fb2d9d142.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 575.523602] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 575.524285] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-86f88d21-cf3d-4a69-b277-921c9ab7a46a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.532251] env[63531]: DEBUG oslo_vmware.api [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Waiting for the task: (returnval){ [ 575.532251] env[63531]: value = "task-1117496" [ 575.532251] env[63531]: _type = "Task" [ 575.532251] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 575.546464] env[63531]: DEBUG oslo_vmware.api [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': task-1117496, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 575.611193] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f3f5aed-688d-44c1-8c59-d4b16408a0bb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.623539] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0cde5bc-04e3-4b9d-a480-64987940f10d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.666851] env[63531]: INFO nova.scheduler.client.report [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Deleted allocations for instance d1636eb1-b862-47c9-8729-7a9d475ace01 [ 575.676191] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f50e04b-4a39-40eb-b90b-542e1cb50e0e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.687311] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b909220-cbe0-41a3-b071-6d9dd434567c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.703533] env[63531]: DEBUG nova.compute.provider_tree [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 575.740638] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 575.829645] env[63531]: DEBUG nova.compute.manager [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 576.045658] env[63531]: DEBUG oslo_vmware.api [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': task-1117496, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.180238} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 576.045920] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 576.046952] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-558dbc20-bf9a-4bd1-ad8e-14aa1f8c5417 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.067237] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Reconfiguring VM instance instance-00000003 to attach disk [datastore2] 4410b013-868a-41e7-9dfb-584fb2d9d142/4410b013-868a-41e7-9dfb-584fb2d9d142.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 576.067449] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-38ed2d9d-2ac2-4e91-aebf-971d8489508c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.086735] env[63531]: DEBUG oslo_vmware.api [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Waiting for the task: (returnval){ [ 576.086735] env[63531]: value = "task-1117497" [ 576.086735] env[63531]: _type = "Task" [ 576.086735] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 576.097849] env[63531]: DEBUG oslo_vmware.api [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': task-1117497, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 576.183134] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e4c0714c-09dc-4f7b-81d8-25539b5dbd81 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Lock "d1636eb1-b862-47c9-8729-7a9d475ace01" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.694s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 576.208826] env[63531]: DEBUG nova.scheduler.client.report [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 576.332724] env[63531]: DEBUG nova.network.neutron [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] Successfully created port: 9608e6f9-9a3e-431c-a11d-02f74c26ee7e {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 576.466958] env[63531]: DEBUG oslo_concurrency.lockutils [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] Acquiring lock "046304ef-b46f-473f-aa5b-5932b5078c54" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 576.466958] env[63531]: DEBUG oslo_concurrency.lockutils [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] Lock "046304ef-b46f-473f-aa5b-5932b5078c54" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 576.598429] env[63531]: DEBUG oslo_vmware.api [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': task-1117497, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 576.686123] env[63531]: DEBUG nova.compute.manager [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 576.720393] env[63531]: DEBUG oslo_concurrency.lockutils [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.902s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 576.722999] env[63531]: ERROR nova.compute.manager [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e70dcd41-edfa-4975-a540-6e7d7bbd6280, please check neutron logs for more information. [ 576.722999] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] Traceback (most recent call last): [ 576.722999] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 576.722999] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] self.driver.spawn(context, instance, image_meta, [ 576.722999] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 576.722999] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] self._vmops.spawn(context, instance, image_meta, injected_files, [ 576.722999] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 576.722999] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] vm_ref = self.build_virtual_machine(instance, [ 576.722999] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 576.722999] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] vif_infos = vmwarevif.get_vif_info(self._session, [ 576.722999] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 576.723574] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] for vif in network_info: [ 576.723574] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 576.723574] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] return self._sync_wrapper(fn, *args, **kwargs) [ 576.723574] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 576.723574] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] self.wait() [ 576.723574] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 576.723574] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] self[:] = self._gt.wait() [ 576.723574] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 576.723574] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] return self._exit_event.wait() [ 576.723574] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 576.723574] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] result = hub.switch() [ 576.723574] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 576.723574] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] return self.greenlet.switch() [ 576.725051] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 576.725051] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] result = function(*args, **kwargs) [ 576.725051] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 576.725051] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] return func(*args, **kwargs) [ 576.725051] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 576.725051] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] raise e [ 576.725051] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 576.725051] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] nwinfo = self.network_api.allocate_for_instance( [ 576.725051] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 576.725051] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] created_port_ids = self._update_ports_for_instance( [ 576.725051] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 576.725051] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] with excutils.save_and_reraise_exception(): [ 576.725051] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 576.725393] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] self.force_reraise() [ 576.725393] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 576.725393] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] raise self.value [ 576.725393] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 576.725393] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] updated_port = self._update_port( [ 576.725393] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 576.725393] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] _ensure_no_port_binding_failure(port) [ 576.725393] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 576.725393] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] raise exception.PortBindingFailed(port_id=port['id']) [ 576.725393] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] nova.exception.PortBindingFailed: Binding failed for port e70dcd41-edfa-4975-a540-6e7d7bbd6280, please check neutron logs for more information. [ 576.725393] env[63531]: ERROR nova.compute.manager [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] [ 576.725674] env[63531]: DEBUG nova.compute.utils [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] Binding failed for port e70dcd41-edfa-4975-a540-6e7d7bbd6280, please check neutron logs for more information. {{(pid=63531) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 576.725674] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 10.572s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 576.728877] env[63531]: DEBUG nova.compute.manager [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] Build of instance 6e963227-64f7-414d-88ec-cee5d5770bba was re-scheduled: Binding failed for port e70dcd41-edfa-4975-a540-6e7d7bbd6280, please check neutron logs for more information. {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 576.733311] env[63531]: DEBUG nova.compute.manager [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] Unplugging VIFs for instance {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 576.735508] env[63531]: DEBUG oslo_concurrency.lockutils [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Acquiring lock "refresh_cache-6e963227-64f7-414d-88ec-cee5d5770bba" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 576.735508] env[63531]: DEBUG oslo_concurrency.lockutils [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Acquired lock "refresh_cache-6e963227-64f7-414d-88ec-cee5d5770bba" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 576.735508] env[63531]: DEBUG nova.network.neutron [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 576.841815] env[63531]: DEBUG nova.compute.manager [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 576.880894] env[63531]: DEBUG nova.virt.hardware [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 576.881187] env[63531]: DEBUG nova.virt.hardware [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 576.881347] env[63531]: DEBUG nova.virt.hardware [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 576.881530] env[63531]: DEBUG nova.virt.hardware [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 576.881676] env[63531]: DEBUG nova.virt.hardware [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 576.882578] env[63531]: DEBUG nova.virt.hardware [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 576.882764] env[63531]: DEBUG nova.virt.hardware [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 576.882932] env[63531]: DEBUG nova.virt.hardware [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 576.883118] env[63531]: DEBUG nova.virt.hardware [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 576.883287] env[63531]: DEBUG nova.virt.hardware [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 576.883461] env[63531]: DEBUG nova.virt.hardware [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 576.884384] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f7d1caf-bf5a-4cb6-93a0-6986fc66374b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.895861] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] Acquiring lock "e1e72ba0-fc50-4776-991f-738c2d9c4ff3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 576.896115] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] Lock "e1e72ba0-fc50-4776-991f-738c2d9c4ff3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 576.902375] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f24e8b2-a997-4030-a5a0-9f0042cd42e9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.102268] env[63531]: DEBUG oslo_vmware.api [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': task-1117497, 'name': ReconfigVM_Task, 'duration_secs': 0.667329} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 577.102268] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Reconfigured VM instance instance-00000003 to attach disk [datastore2] 4410b013-868a-41e7-9dfb-584fb2d9d142/4410b013-868a-41e7-9dfb-584fb2d9d142.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 577.102268] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-032d0841-f9a3-4dc2-b83e-a7dcc25cbb76 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.116223] env[63531]: DEBUG oslo_vmware.api [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Waiting for the task: (returnval){ [ 577.116223] env[63531]: value = "task-1117498" [ 577.116223] env[63531]: _type = "Task" [ 577.116223] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 577.125194] env[63531]: DEBUG oslo_vmware.api [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': task-1117498, 'name': Rename_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 577.214900] env[63531]: DEBUG oslo_concurrency.lockutils [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 577.295317] env[63531]: DEBUG nova.network.neutron [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 577.531991] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcb98df4-6a35-46c0-af8f-f7e5a453edfa {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.544370] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53a6f7a5-c4e3-4896-bab6-7eeb42c5f5ad {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.579723] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2d92cc2-fcfe-474c-b895-d3b657018770 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.588111] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0485568-2c8b-4d87-8d4c-c38b30437343 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.602273] env[63531]: DEBUG nova.compute.provider_tree [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 577.625603] env[63531]: DEBUG oslo_vmware.api [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': task-1117498, 'name': Rename_Task, 'duration_secs': 0.204308} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 577.626762] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 577.626762] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2ab844f1-a616-42fc-a387-185d6c89476f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.633212] env[63531]: DEBUG oslo_vmware.api [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Waiting for the task: (returnval){ [ 577.633212] env[63531]: value = "task-1117499" [ 577.633212] env[63531]: _type = "Task" [ 577.633212] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 577.643982] env[63531]: DEBUG oslo_vmware.api [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': task-1117499, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 577.659538] env[63531]: DEBUG nova.network.neutron [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 578.106467] env[63531]: DEBUG nova.scheduler.client.report [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 578.151312] env[63531]: DEBUG oslo_vmware.api [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': task-1117499, 'name': PowerOnVM_Task} progress is 84%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 578.163895] env[63531]: DEBUG oslo_concurrency.lockutils [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Releasing lock "refresh_cache-6e963227-64f7-414d-88ec-cee5d5770bba" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 578.163895] env[63531]: DEBUG nova.compute.manager [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 578.163895] env[63531]: DEBUG nova.compute.manager [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 578.163895] env[63531]: DEBUG nova.network.neutron [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 578.195280] env[63531]: DEBUG nova.network.neutron [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 578.615338] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.892s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 578.616650] env[63531]: ERROR nova.compute.manager [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e006227b-9069-49b7-a244-083d6aed19d7, please check neutron logs for more information. [ 578.616650] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] Traceback (most recent call last): [ 578.616650] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 578.616650] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] self.driver.spawn(context, instance, image_meta, [ 578.616650] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 578.616650] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] self._vmops.spawn(context, instance, image_meta, injected_files, [ 578.616650] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 578.616650] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] vm_ref = self.build_virtual_machine(instance, [ 578.616650] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 578.616650] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] vif_infos = vmwarevif.get_vif_info(self._session, [ 578.616650] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 578.617064] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] for vif in network_info: [ 578.617064] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 578.617064] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] return self._sync_wrapper(fn, *args, **kwargs) [ 578.617064] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 578.617064] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] self.wait() [ 578.617064] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 578.617064] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] self[:] = self._gt.wait() [ 578.617064] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 578.617064] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] return self._exit_event.wait() [ 578.617064] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 578.617064] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] result = hub.switch() [ 578.617064] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 578.617064] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] return self.greenlet.switch() [ 578.617448] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 578.617448] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] result = function(*args, **kwargs) [ 578.617448] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 578.617448] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] return func(*args, **kwargs) [ 578.617448] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 578.617448] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] raise e [ 578.617448] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 578.617448] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] nwinfo = self.network_api.allocate_for_instance( [ 578.617448] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 578.617448] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] created_port_ids = self._update_ports_for_instance( [ 578.617448] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 578.617448] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] with excutils.save_and_reraise_exception(): [ 578.617448] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 578.617871] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] self.force_reraise() [ 578.617871] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 578.617871] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] raise self.value [ 578.617871] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 578.617871] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] updated_port = self._update_port( [ 578.617871] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 578.617871] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] _ensure_no_port_binding_failure(port) [ 578.617871] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 578.617871] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] raise exception.PortBindingFailed(port_id=port['id']) [ 578.617871] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] nova.exception.PortBindingFailed: Binding failed for port e006227b-9069-49b7-a244-083d6aed19d7, please check neutron logs for more information. [ 578.617871] env[63531]: ERROR nova.compute.manager [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] [ 578.618167] env[63531]: DEBUG nova.compute.utils [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] Binding failed for port e006227b-9069-49b7-a244-083d6aed19d7, please check neutron logs for more information. {{(pid=63531) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 578.620729] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 10.609s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 578.625246] env[63531]: DEBUG nova.compute.manager [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] Build of instance 35e9c404-bad4-4f8d-8c79-fb7250abb363 was re-scheduled: Binding failed for port e006227b-9069-49b7-a244-083d6aed19d7, please check neutron logs for more information. {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 578.625724] env[63531]: DEBUG nova.compute.manager [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] Unplugging VIFs for instance {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 578.625833] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] Acquiring lock "refresh_cache-35e9c404-bad4-4f8d-8c79-fb7250abb363" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 578.626114] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] Acquired lock "refresh_cache-35e9c404-bad4-4f8d-8c79-fb7250abb363" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 578.626114] env[63531]: DEBUG nova.network.neutron [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 578.651371] env[63531]: DEBUG oslo_vmware.api [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': task-1117499, 'name': PowerOnVM_Task, 'duration_secs': 0.810793} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 578.652403] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 578.652403] env[63531]: DEBUG nova.compute.manager [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 578.653106] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c22bc66e-a963-49df-a146-5e8bc6524b2d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.701503] env[63531]: DEBUG nova.network.neutron [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 579.004775] env[63531]: ERROR nova.compute.manager [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9608e6f9-9a3e-431c-a11d-02f74c26ee7e, please check neutron logs for more information. [ 579.004775] env[63531]: ERROR nova.compute.manager Traceback (most recent call last): [ 579.004775] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 579.004775] env[63531]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 579.004775] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 579.004775] env[63531]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 579.004775] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 579.004775] env[63531]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 579.004775] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 579.004775] env[63531]: ERROR nova.compute.manager self.force_reraise() [ 579.004775] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 579.004775] env[63531]: ERROR nova.compute.manager raise self.value [ 579.004775] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 579.004775] env[63531]: ERROR nova.compute.manager updated_port = self._update_port( [ 579.004775] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 579.004775] env[63531]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 579.005339] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 579.005339] env[63531]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 579.005339] env[63531]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9608e6f9-9a3e-431c-a11d-02f74c26ee7e, please check neutron logs for more information. [ 579.005339] env[63531]: ERROR nova.compute.manager [ 579.005339] env[63531]: Traceback (most recent call last): [ 579.005339] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 579.005339] env[63531]: listener.cb(fileno) [ 579.005339] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 579.005339] env[63531]: result = function(*args, **kwargs) [ 579.005339] env[63531]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 579.005339] env[63531]: return func(*args, **kwargs) [ 579.005339] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 579.005339] env[63531]: raise e [ 579.005339] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 579.005339] env[63531]: nwinfo = self.network_api.allocate_for_instance( [ 579.005339] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 579.005339] env[63531]: created_port_ids = self._update_ports_for_instance( [ 579.005339] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 579.005339] env[63531]: with excutils.save_and_reraise_exception(): [ 579.005339] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 579.005339] env[63531]: self.force_reraise() [ 579.005339] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 579.005339] env[63531]: raise self.value [ 579.005339] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 579.005339] env[63531]: updated_port = self._update_port( [ 579.005339] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 579.005339] env[63531]: _ensure_no_port_binding_failure(port) [ 579.005339] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 579.005339] env[63531]: raise exception.PortBindingFailed(port_id=port['id']) [ 579.006162] env[63531]: nova.exception.PortBindingFailed: Binding failed for port 9608e6f9-9a3e-431c-a11d-02f74c26ee7e, please check neutron logs for more information. [ 579.006162] env[63531]: Removing descriptor: 17 [ 579.006162] env[63531]: ERROR nova.compute.manager [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9608e6f9-9a3e-431c-a11d-02f74c26ee7e, please check neutron logs for more information. [ 579.006162] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] Traceback (most recent call last): [ 579.006162] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 579.006162] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] yield resources [ 579.006162] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 579.006162] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] self.driver.spawn(context, instance, image_meta, [ 579.006162] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 579.006162] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 579.006162] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 579.006162] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] vm_ref = self.build_virtual_machine(instance, [ 579.006525] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 579.006525] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] vif_infos = vmwarevif.get_vif_info(self._session, [ 579.006525] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 579.006525] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] for vif in network_info: [ 579.006525] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 579.006525] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] return self._sync_wrapper(fn, *args, **kwargs) [ 579.006525] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 579.006525] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] self.wait() [ 579.006525] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 579.006525] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] self[:] = self._gt.wait() [ 579.006525] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 579.006525] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] return self._exit_event.wait() [ 579.006525] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 579.006918] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] result = hub.switch() [ 579.006918] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 579.006918] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] return self.greenlet.switch() [ 579.006918] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 579.006918] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] result = function(*args, **kwargs) [ 579.006918] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 579.006918] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] return func(*args, **kwargs) [ 579.006918] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 579.006918] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] raise e [ 579.006918] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 579.006918] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] nwinfo = self.network_api.allocate_for_instance( [ 579.006918] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 579.006918] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] created_port_ids = self._update_ports_for_instance( [ 579.007372] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 579.007372] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] with excutils.save_and_reraise_exception(): [ 579.007372] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 579.007372] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] self.force_reraise() [ 579.007372] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 579.007372] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] raise self.value [ 579.007372] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 579.007372] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] updated_port = self._update_port( [ 579.007372] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 579.007372] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] _ensure_no_port_binding_failure(port) [ 579.007372] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 579.007372] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] raise exception.PortBindingFailed(port_id=port['id']) [ 579.007722] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] nova.exception.PortBindingFailed: Binding failed for port 9608e6f9-9a3e-431c-a11d-02f74c26ee7e, please check neutron logs for more information. [ 579.007722] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] [ 579.007722] env[63531]: INFO nova.compute.manager [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] Terminating instance [ 579.010648] env[63531]: DEBUG oslo_concurrency.lockutils [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] Acquiring lock "refresh_cache-2ae517ae-a159-4dc6-8100-457a4cea1aeb" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 579.010648] env[63531]: DEBUG oslo_concurrency.lockutils [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] Acquired lock "refresh_cache-2ae517ae-a159-4dc6-8100-457a4cea1aeb" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 579.010648] env[63531]: DEBUG nova.network.neutron [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 579.171884] env[63531]: DEBUG nova.network.neutron [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 579.177022] env[63531]: DEBUG oslo_concurrency.lockutils [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 579.214487] env[63531]: INFO nova.compute.manager [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 6e963227-64f7-414d-88ec-cee5d5770bba] Took 1.05 seconds to deallocate network for instance. [ 579.472442] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37aa4b6c-3db2-4139-89ba-a6823d0a001d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.480163] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e98d6b23-5bf5-45f7-a23f-4bf767d38e6d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.516100] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d46fa644-4253-4d1f-bbd7-550eb351be3a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.525022] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c893399-6358-4f80-8f6d-24ece449ff3f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.543437] env[63531]: DEBUG nova.compute.provider_tree [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 579.664089] env[63531]: DEBUG nova.compute.manager [req-87fd27d1-4731-4c65-9c8d-e379d0186983 req-30813bba-1927-42ad-bd82-c0ae2ba2ad87 service nova] [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] Received event network-changed-9608e6f9-9a3e-431c-a11d-02f74c26ee7e {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 579.664412] env[63531]: DEBUG nova.compute.manager [req-87fd27d1-4731-4c65-9c8d-e379d0186983 req-30813bba-1927-42ad-bd82-c0ae2ba2ad87 service nova] [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] Refreshing instance network info cache due to event network-changed-9608e6f9-9a3e-431c-a11d-02f74c26ee7e. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 579.664568] env[63531]: DEBUG oslo_concurrency.lockutils [req-87fd27d1-4731-4c65-9c8d-e379d0186983 req-30813bba-1927-42ad-bd82-c0ae2ba2ad87 service nova] Acquiring lock "refresh_cache-2ae517ae-a159-4dc6-8100-457a4cea1aeb" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 579.716053] env[63531]: DEBUG nova.network.neutron [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 579.874702] env[63531]: DEBUG nova.network.neutron [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 579.984202] env[63531]: DEBUG nova.network.neutron [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 580.046696] env[63531]: DEBUG nova.scheduler.client.report [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 580.265229] env[63531]: INFO nova.scheduler.client.report [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Deleted allocations for instance 6e963227-64f7-414d-88ec-cee5d5770bba [ 580.324537] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Acquiring lock "316c12f7-bc05-4f70-98b4-579d4a7b69d0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.324804] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Lock "316c12f7-bc05-4f70-98b4-579d4a7b69d0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 580.377799] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] Releasing lock "refresh_cache-35e9c404-bad4-4f8d-8c79-fb7250abb363" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 580.379948] env[63531]: DEBUG nova.compute.manager [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 580.379948] env[63531]: DEBUG nova.compute.manager [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 580.379948] env[63531]: DEBUG nova.network.neutron [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 580.418668] env[63531]: DEBUG nova.network.neutron [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 580.487403] env[63531]: DEBUG oslo_concurrency.lockutils [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] Releasing lock "refresh_cache-2ae517ae-a159-4dc6-8100-457a4cea1aeb" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 580.488219] env[63531]: DEBUG nova.compute.manager [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 580.488219] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 580.488435] env[63531]: DEBUG oslo_concurrency.lockutils [req-87fd27d1-4731-4c65-9c8d-e379d0186983 req-30813bba-1927-42ad-bd82-c0ae2ba2ad87 service nova] Acquired lock "refresh_cache-2ae517ae-a159-4dc6-8100-457a4cea1aeb" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 580.488630] env[63531]: DEBUG nova.network.neutron [req-87fd27d1-4731-4c65-9c8d-e379d0186983 req-30813bba-1927-42ad-bd82-c0ae2ba2ad87 service nova] [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] Refreshing network info cache for port 9608e6f9-9a3e-431c-a11d-02f74c26ee7e {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 580.493812] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c10f362a-bb01-4777-900a-31e8bd360b34 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.503193] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ae032d2-31db-46be-9e79-4f74a64228e2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.530848] env[63531]: WARNING nova.virt.vmwareapi.vmops [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2ae517ae-a159-4dc6-8100-457a4cea1aeb could not be found. [ 580.531234] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 580.531282] env[63531]: INFO nova.compute.manager [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] Took 0.04 seconds to destroy the instance on the hypervisor. [ 580.531513] env[63531]: DEBUG oslo.service.loopingcall [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 580.531716] env[63531]: DEBUG nova.compute.manager [-] [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 580.531810] env[63531]: DEBUG nova.network.neutron [-] [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 580.556443] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.931s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 580.556443] env[63531]: ERROR nova.compute.manager [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 60020572-c02a-4101-9cec-88afc59a70d9, please check neutron logs for more information. [ 580.556443] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] Traceback (most recent call last): [ 580.556443] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 580.556443] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] self.driver.spawn(context, instance, image_meta, [ 580.556443] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 580.556443] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 580.556443] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 580.556443] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] vm_ref = self.build_virtual_machine(instance, [ 580.556789] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 580.556789] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] vif_infos = vmwarevif.get_vif_info(self._session, [ 580.556789] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 580.556789] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] for vif in network_info: [ 580.556789] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 580.556789] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] return self._sync_wrapper(fn, *args, **kwargs) [ 580.556789] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 580.556789] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] self.wait() [ 580.556789] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 580.556789] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] self[:] = self._gt.wait() [ 580.556789] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 580.556789] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] return self._exit_event.wait() [ 580.556789] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 580.557177] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] result = hub.switch() [ 580.557177] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 580.557177] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] return self.greenlet.switch() [ 580.557177] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 580.557177] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] result = function(*args, **kwargs) [ 580.557177] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 580.557177] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] return func(*args, **kwargs) [ 580.557177] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 580.557177] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] raise e [ 580.557177] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 580.557177] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] nwinfo = self.network_api.allocate_for_instance( [ 580.557177] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 580.557177] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] created_port_ids = self._update_ports_for_instance( [ 580.557538] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 580.557538] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] with excutils.save_and_reraise_exception(): [ 580.557538] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 580.557538] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] self.force_reraise() [ 580.557538] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 580.557538] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] raise self.value [ 580.557538] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 580.557538] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] updated_port = self._update_port( [ 580.557538] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 580.557538] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] _ensure_no_port_binding_failure(port) [ 580.557538] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 580.557538] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] raise exception.PortBindingFailed(port_id=port['id']) [ 580.557854] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] nova.exception.PortBindingFailed: Binding failed for port 60020572-c02a-4101-9cec-88afc59a70d9, please check neutron logs for more information. [ 580.557854] env[63531]: ERROR nova.compute.manager [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] [ 580.557854] env[63531]: DEBUG nova.compute.utils [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] Binding failed for port 60020572-c02a-4101-9cec-88afc59a70d9, please check neutron logs for more information. {{(pid=63531) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 580.557854] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 11.586s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 580.557854] env[63531]: DEBUG nova.compute.manager [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] Build of instance 5287b3ed-12ec-47fb-93a2-94038cb2e7b1 was re-scheduled: Binding failed for port 60020572-c02a-4101-9cec-88afc59a70d9, please check neutron logs for more information. {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 580.558042] env[63531]: DEBUG nova.compute.manager [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] Unplugging VIFs for instance {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 580.558042] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] Acquiring lock "refresh_cache-5287b3ed-12ec-47fb-93a2-94038cb2e7b1" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 580.558111] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] Acquired lock "refresh_cache-5287b3ed-12ec-47fb-93a2-94038cb2e7b1" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 580.558942] env[63531]: DEBUG nova.network.neutron [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 580.579974] env[63531]: DEBUG nova.network.neutron [-] [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 580.779468] env[63531]: DEBUG oslo_concurrency.lockutils [None req-29264508-166c-47d2-ba89-a1a3a137bc6b tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Lock "6e963227-64f7-414d-88ec-cee5d5770bba" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.602s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 580.928682] env[63531]: DEBUG nova.network.neutron [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 581.036938] env[63531]: DEBUG nova.network.neutron [req-87fd27d1-4731-4c65-9c8d-e379d0186983 req-30813bba-1927-42ad-bd82-c0ae2ba2ad87 service nova] [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 581.081789] env[63531]: DEBUG nova.network.neutron [-] [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 581.108765] env[63531]: DEBUG nova.network.neutron [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 581.278618] env[63531]: DEBUG nova.network.neutron [req-87fd27d1-4731-4c65-9c8d-e379d0186983 req-30813bba-1927-42ad-bd82-c0ae2ba2ad87 service nova] [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 581.283421] env[63531]: DEBUG nova.compute.manager [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 581.406172] env[63531]: DEBUG nova.network.neutron [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 581.435392] env[63531]: INFO nova.compute.manager [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] [instance: 35e9c404-bad4-4f8d-8c79-fb7250abb363] Took 1.05 seconds to deallocate network for instance. [ 581.584630] env[63531]: INFO nova.compute.manager [-] [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] Took 1.05 seconds to deallocate network for instance. [ 581.590299] env[63531]: DEBUG nova.compute.claims [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] Aborting claim: {{(pid=63531) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 581.591136] env[63531]: DEBUG oslo_concurrency.lockutils [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 581.607345] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 4410b013-868a-41e7-9dfb-584fb2d9d142 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 581.782150] env[63531]: DEBUG oslo_concurrency.lockutils [req-87fd27d1-4731-4c65-9c8d-e379d0186983 req-30813bba-1927-42ad-bd82-c0ae2ba2ad87 service nova] Releasing lock "refresh_cache-2ae517ae-a159-4dc6-8100-457a4cea1aeb" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 581.811702] env[63531]: DEBUG oslo_concurrency.lockutils [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 581.917410] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] Releasing lock "refresh_cache-5287b3ed-12ec-47fb-93a2-94038cb2e7b1" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 581.917410] env[63531]: DEBUG nova.compute.manager [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 581.917410] env[63531]: DEBUG nova.compute.manager [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 581.917410] env[63531]: DEBUG nova.network.neutron [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 581.945014] env[63531]: DEBUG nova.network.neutron [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 582.023599] env[63531]: INFO nova.compute.manager [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Rebuilding instance [ 582.086934] env[63531]: DEBUG nova.compute.manager [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 582.087891] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d390f628-0c6f-4344-bbb5-3463d23ce82b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.111134] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 35e9c404-bad4-4f8d-8c79-fb7250abb363 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 582.219444] env[63531]: DEBUG nova.compute.manager [req-6784f978-aaf6-4e6d-a484-e5997d1c27d5 req-d0133c5d-84d8-4259-82c2-ad595fa41de1 service nova] [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] Received event network-vif-deleted-9608e6f9-9a3e-431c-a11d-02f74c26ee7e {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 582.454016] env[63531]: DEBUG nova.network.neutron [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 582.477828] env[63531]: INFO nova.scheduler.client.report [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] Deleted allocations for instance 35e9c404-bad4-4f8d-8c79-fb7250abb363 [ 582.604875] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 582.604875] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-262ddcf3-9372-4427-8ec6-e3f828afe08a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.612691] env[63531]: DEBUG oslo_vmware.api [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Waiting for the task: (returnval){ [ 582.612691] env[63531]: value = "task-1117502" [ 582.612691] env[63531]: _type = "Task" [ 582.612691] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 582.619404] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 5287b3ed-12ec-47fb-93a2-94038cb2e7b1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 582.620369] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 5b86b63a-f6c1-4983-a3cc-3cde7a10d17b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 582.620369] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 6a380967-bca6-47d2-b052-988e6c1004ee actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 582.620369] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 2ae517ae-a159-4dc6-8100-457a4cea1aeb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 582.632778] env[63531]: DEBUG oslo_vmware.api [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Task: {'id': task-1117502, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 582.962959] env[63531]: INFO nova.compute.manager [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] [instance: 5287b3ed-12ec-47fb-93a2-94038cb2e7b1] Took 1.04 seconds to deallocate network for instance. [ 582.987806] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bf6f74b5-db8c-4a10-8223-8ce91114afd6 tempest-ImagesOneServerNegativeTestJSON-913282109 tempest-ImagesOneServerNegativeTestJSON-913282109-project-member] Lock "35e9c404-bad4-4f8d-8c79-fb7250abb363" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.224s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 583.129233] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 9fa373b3-8e15-4d26-82be-7dda51bc6cf6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 583.137873] env[63531]: DEBUG oslo_vmware.api [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Task: {'id': task-1117502, 'name': PowerOffVM_Task, 'duration_secs': 0.118004} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 583.141032] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 583.141032] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 583.141032] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e1a6b5f-86d1-4a8c-8e3a-7fb5d5ccff00 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.150269] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 583.150679] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cb75112b-3bf2-4281-aff1-4296c8ad5aa0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.178208] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 583.178455] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Deleting contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 583.178662] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Deleting the datastore file [datastore2] 4410b013-868a-41e7-9dfb-584fb2d9d142 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 583.178953] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ce5e3bb1-ec1d-4117-9d31-b121cbbfc37c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.185733] env[63531]: DEBUG oslo_vmware.api [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Waiting for the task: (returnval){ [ 583.185733] env[63531]: value = "task-1117505" [ 583.185733] env[63531]: _type = "Task" [ 583.185733] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 583.193528] env[63531]: DEBUG oslo_vmware.api [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Task: {'id': task-1117505, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 583.496297] env[63531]: DEBUG nova.compute.manager [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 583.641504] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance d941a593-7047-436a-b95d-bcb4850829f5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 583.645940] env[63531]: DEBUG nova.compute.manager [None req-fe82ffe7-25fb-4e88-9cc8-215aba4f1342 tempest-ServerDiagnosticsV248Test-1489378919 tempest-ServerDiagnosticsV248Test-1489378919-project-admin] [instance: 5b86b63a-f6c1-4983-a3cc-3cde7a10d17b] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 583.648069] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87c03b20-00d5-4d56-bafd-74a4317ae9ed {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.660467] env[63531]: INFO nova.compute.manager [None req-fe82ffe7-25fb-4e88-9cc8-215aba4f1342 tempest-ServerDiagnosticsV248Test-1489378919 tempest-ServerDiagnosticsV248Test-1489378919-project-admin] [instance: 5b86b63a-f6c1-4983-a3cc-3cde7a10d17b] Retrieving diagnostics [ 583.661851] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-304e331f-e017-4523-9eab-367e9f38df72 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.710388] env[63531]: DEBUG oslo_vmware.api [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Task: {'id': task-1117505, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.113588} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 583.710668] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 583.710940] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Deleted contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 583.711071] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 583.961814] env[63531]: DEBUG oslo_concurrency.lockutils [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] Acquiring lock "180526ef-0e57-4834-bc35-1e8e086e7323" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 583.962074] env[63531]: DEBUG oslo_concurrency.lockutils [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] Lock "180526ef-0e57-4834-bc35-1e8e086e7323" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 584.001177] env[63531]: INFO nova.scheduler.client.report [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] Deleted allocations for instance 5287b3ed-12ec-47fb-93a2-94038cb2e7b1 [ 584.038549] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 584.150053] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 2b5e3059-2772-496b-95c7-a083a37ef155 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 584.521256] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f35c576d-c593-4d9f-b00b-ea633565c2e4 tempest-ImagesOneServerTestJSON-290929020 tempest-ImagesOneServerTestJSON-290929020-project-member] Lock "5287b3ed-12ec-47fb-93a2-94038cb2e7b1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.702s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 584.655387] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 01a1b3e0-4ad9-4350-b59e-499f1b3412a5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 584.760489] env[63531]: DEBUG nova.virt.hardware [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 584.760770] env[63531]: DEBUG nova.virt.hardware [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 584.760968] env[63531]: DEBUG nova.virt.hardware [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 584.761173] env[63531]: DEBUG nova.virt.hardware [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 584.761384] env[63531]: DEBUG nova.virt.hardware [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 584.761546] env[63531]: DEBUG nova.virt.hardware [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 584.762507] env[63531]: DEBUG nova.virt.hardware [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 584.763412] env[63531]: DEBUG nova.virt.hardware [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 584.763412] env[63531]: DEBUG nova.virt.hardware [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 584.763412] env[63531]: DEBUG nova.virt.hardware [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 584.763412] env[63531]: DEBUG nova.virt.hardware [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 584.764210] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb97d135-a5b0-4d04-9dd0-d98cdd0d2c72 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.773383] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f544cdff-a87b-498a-a729-769bb37cd4e4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.790614] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Instance VIF info [] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 584.799590] env[63531]: DEBUG oslo.service.loopingcall [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 584.800015] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 584.800113] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-40ce4b81-8011-490e-a23f-5eac136e4e5d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.817362] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 584.817362] env[63531]: value = "task-1117506" [ 584.817362] env[63531]: _type = "Task" [ 584.817362] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 584.826035] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117506, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 585.023989] env[63531]: DEBUG nova.compute.manager [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 585.162708] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 8ac0012a-0855-4b73-a1e4-0a68a1336404 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 585.329794] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117506, 'name': CreateVM_Task, 'duration_secs': 0.378144} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 585.331786] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 585.332155] env[63531]: DEBUG oslo_concurrency.lockutils [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 585.332648] env[63531]: DEBUG oslo_concurrency.lockutils [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 585.333278] env[63531]: DEBUG oslo_concurrency.lockutils [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 585.335161] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3143209e-818b-4e99-ab74-9861de71b4e4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.347694] env[63531]: DEBUG oslo_vmware.api [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Waiting for the task: (returnval){ [ 585.347694] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52beb860-de6e-6fa8-8351-bfc7765b3b50" [ 585.347694] env[63531]: _type = "Task" [ 585.347694] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 585.357520] env[63531]: DEBUG oslo_vmware.api [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52beb860-de6e-6fa8-8351-bfc7765b3b50, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 585.556835] env[63531]: DEBUG oslo_concurrency.lockutils [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 585.669817] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 3a02946a-64e9-41b7-95d7-0b32766d735c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 585.783177] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a8d3fe75-9fbe-4167-b227-285d91eb9660 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Acquiring lock "5b86b63a-f6c1-4983-a3cc-3cde7a10d17b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 585.783177] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a8d3fe75-9fbe-4167-b227-285d91eb9660 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Lock "5b86b63a-f6c1-4983-a3cc-3cde7a10d17b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 585.783177] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a8d3fe75-9fbe-4167-b227-285d91eb9660 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Acquiring lock "5b86b63a-f6c1-4983-a3cc-3cde7a10d17b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 585.783177] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a8d3fe75-9fbe-4167-b227-285d91eb9660 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Lock "5b86b63a-f6c1-4983-a3cc-3cde7a10d17b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 585.783474] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a8d3fe75-9fbe-4167-b227-285d91eb9660 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Lock "5b86b63a-f6c1-4983-a3cc-3cde7a10d17b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 585.785335] env[63531]: INFO nova.compute.manager [None req-a8d3fe75-9fbe-4167-b227-285d91eb9660 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] [instance: 5b86b63a-f6c1-4983-a3cc-3cde7a10d17b] Terminating instance [ 585.787222] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a8d3fe75-9fbe-4167-b227-285d91eb9660 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Acquiring lock "refresh_cache-5b86b63a-f6c1-4983-a3cc-3cde7a10d17b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 585.787510] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a8d3fe75-9fbe-4167-b227-285d91eb9660 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Acquired lock "refresh_cache-5b86b63a-f6c1-4983-a3cc-3cde7a10d17b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 585.788079] env[63531]: DEBUG nova.network.neutron [None req-a8d3fe75-9fbe-4167-b227-285d91eb9660 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] [instance: 5b86b63a-f6c1-4983-a3cc-3cde7a10d17b] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 585.859365] env[63531]: DEBUG oslo_vmware.api [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52beb860-de6e-6fa8-8351-bfc7765b3b50, 'name': SearchDatastore_Task, 'duration_secs': 0.010516} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 585.859600] env[63531]: DEBUG oslo_concurrency.lockutils [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 585.859943] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 585.860072] env[63531]: DEBUG oslo_concurrency.lockutils [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 585.860224] env[63531]: DEBUG oslo_concurrency.lockutils [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 585.860617] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 585.860831] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3baabeb0-ccdb-4e77-bcf6-28557088e242 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.870112] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 585.870659] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 585.872262] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6606acba-6d42-44c5-9539-5fa6834f7a4a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.880881] env[63531]: DEBUG oslo_vmware.api [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Waiting for the task: (returnval){ [ 585.880881] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52421368-5270-4f8b-5732-9c413cd9d0a5" [ 585.880881] env[63531]: _type = "Task" [ 585.880881] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 585.892429] env[63531]: DEBUG oslo_vmware.api [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52421368-5270-4f8b-5732-9c413cd9d0a5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 586.178902] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance fe881dbe-6d44-4e31-a537-4aa9dfe44ff4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 586.337796] env[63531]: DEBUG nova.network.neutron [None req-a8d3fe75-9fbe-4167-b227-285d91eb9660 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] [instance: 5b86b63a-f6c1-4983-a3cc-3cde7a10d17b] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 586.396920] env[63531]: DEBUG oslo_vmware.api [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52421368-5270-4f8b-5732-9c413cd9d0a5, 'name': SearchDatastore_Task, 'duration_secs': 0.009852} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 586.398157] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-32232924-9b2d-41fd-9b8b-5e0723a6f2b9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.404142] env[63531]: DEBUG oslo_vmware.api [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Waiting for the task: (returnval){ [ 586.404142] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]522742be-f83d-77f5-4928-eb1557e06503" [ 586.404142] env[63531]: _type = "Task" [ 586.404142] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 586.415398] env[63531]: DEBUG oslo_vmware.api [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]522742be-f83d-77f5-4928-eb1557e06503, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 586.592212] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Acquiring lock "beb4b334-7d87-4c25-8efe-df92c43f0d26" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 586.592212] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Lock "beb4b334-7d87-4c25-8efe-df92c43f0d26" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 586.657097] env[63531]: DEBUG nova.network.neutron [None req-a8d3fe75-9fbe-4167-b227-285d91eb9660 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] [instance: 5b86b63a-f6c1-4983-a3cc-3cde7a10d17b] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 586.682762] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 3c513284-4860-46b2-8ba1-e185efe0db86 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 586.919044] env[63531]: DEBUG oslo_vmware.api [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]522742be-f83d-77f5-4928-eb1557e06503, 'name': SearchDatastore_Task, 'duration_secs': 0.013281} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 586.919323] env[63531]: DEBUG oslo_concurrency.lockutils [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 586.919583] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 4410b013-868a-41e7-9dfb-584fb2d9d142/4410b013-868a-41e7-9dfb-584fb2d9d142.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 586.919843] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-152c198f-c9cb-479b-82ef-ada01ac64bcd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.929499] env[63531]: DEBUG oslo_vmware.api [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Waiting for the task: (returnval){ [ 586.929499] env[63531]: value = "task-1117507" [ 586.929499] env[63531]: _type = "Task" [ 586.929499] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 586.940129] env[63531]: DEBUG oslo_vmware.api [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Task: {'id': task-1117507, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 587.160850] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a8d3fe75-9fbe-4167-b227-285d91eb9660 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Releasing lock "refresh_cache-5b86b63a-f6c1-4983-a3cc-3cde7a10d17b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 587.161282] env[63531]: DEBUG nova.compute.manager [None req-a8d3fe75-9fbe-4167-b227-285d91eb9660 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] [instance: 5b86b63a-f6c1-4983-a3cc-3cde7a10d17b] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 587.161503] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a8d3fe75-9fbe-4167-b227-285d91eb9660 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] [instance: 5b86b63a-f6c1-4983-a3cc-3cde7a10d17b] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 587.162489] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dc9ae95-a1e9-42b6-9e8d-4b5dd701315c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.179037] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8d3fe75-9fbe-4167-b227-285d91eb9660 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] [instance: 5b86b63a-f6c1-4983-a3cc-3cde7a10d17b] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 587.179571] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-87dad03c-0b80-4fe7-8e98-845858ded6ba {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.185635] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 56872ede-96f0-48ca-b289-86784f37527d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 587.192152] env[63531]: DEBUG oslo_vmware.api [None req-a8d3fe75-9fbe-4167-b227-285d91eb9660 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Waiting for the task: (returnval){ [ 587.192152] env[63531]: value = "task-1117508" [ 587.192152] env[63531]: _type = "Task" [ 587.192152] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 587.204422] env[63531]: DEBUG oslo_vmware.api [None req-a8d3fe75-9fbe-4167-b227-285d91eb9660 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Task: {'id': task-1117508, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 587.448136] env[63531]: DEBUG oslo_vmware.api [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Task: {'id': task-1117507, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.515523} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 587.449304] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 4410b013-868a-41e7-9dfb-584fb2d9d142/4410b013-868a-41e7-9dfb-584fb2d9d142.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 587.449304] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 587.449304] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fbeb9c4e-f7f3-4fbc-9704-4fc690838095 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.457621] env[63531]: DEBUG oslo_vmware.api [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Waiting for the task: (returnval){ [ 587.457621] env[63531]: value = "task-1117509" [ 587.457621] env[63531]: _type = "Task" [ 587.457621] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 587.468920] env[63531]: DEBUG oslo_vmware.api [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Task: {'id': task-1117509, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 587.692600] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 046304ef-b46f-473f-aa5b-5932b5078c54 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 587.709030] env[63531]: DEBUG oslo_vmware.api [None req-a8d3fe75-9fbe-4167-b227-285d91eb9660 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Task: {'id': task-1117508, 'name': PowerOffVM_Task, 'duration_secs': 0.15679} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 587.709030] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8d3fe75-9fbe-4167-b227-285d91eb9660 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] [instance: 5b86b63a-f6c1-4983-a3cc-3cde7a10d17b] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 587.709030] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a8d3fe75-9fbe-4167-b227-285d91eb9660 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] [instance: 5b86b63a-f6c1-4983-a3cc-3cde7a10d17b] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 587.709030] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3cac7a07-bf39-484d-bccd-e25f2c00dc67 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.735365] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a8d3fe75-9fbe-4167-b227-285d91eb9660 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] [instance: 5b86b63a-f6c1-4983-a3cc-3cde7a10d17b] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 587.735490] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a8d3fe75-9fbe-4167-b227-285d91eb9660 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] [instance: 5b86b63a-f6c1-4983-a3cc-3cde7a10d17b] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 587.736461] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8d3fe75-9fbe-4167-b227-285d91eb9660 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Deleting the datastore file [datastore1] 5b86b63a-f6c1-4983-a3cc-3cde7a10d17b {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 587.736461] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-94d886c4-1c1a-4d06-b46b-c6e41f48323c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.742962] env[63531]: DEBUG oslo_vmware.api [None req-a8d3fe75-9fbe-4167-b227-285d91eb9660 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Waiting for the task: (returnval){ [ 587.742962] env[63531]: value = "task-1117511" [ 587.742962] env[63531]: _type = "Task" [ 587.742962] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 587.752246] env[63531]: DEBUG oslo_vmware.api [None req-a8d3fe75-9fbe-4167-b227-285d91eb9660 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Task: {'id': task-1117511, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 587.971150] env[63531]: DEBUG oslo_vmware.api [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Task: {'id': task-1117509, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069436} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 587.971519] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 587.972155] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18f64174-8878-4e18-b0a0-360ef503f77a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.992179] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Reconfiguring VM instance instance-00000003 to attach disk [datastore1] 4410b013-868a-41e7-9dfb-584fb2d9d142/4410b013-868a-41e7-9dfb-584fb2d9d142.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 587.992476] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6411579e-ebf4-43c1-aca1-bb920776be51 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.012746] env[63531]: DEBUG oslo_vmware.api [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Waiting for the task: (returnval){ [ 588.012746] env[63531]: value = "task-1117512" [ 588.012746] env[63531]: _type = "Task" [ 588.012746] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 588.022053] env[63531]: DEBUG oslo_vmware.api [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Task: {'id': task-1117512, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 588.203524] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance e1e72ba0-fc50-4776-991f-738c2d9c4ff3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 588.251732] env[63531]: DEBUG oslo_vmware.api [None req-a8d3fe75-9fbe-4167-b227-285d91eb9660 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Task: {'id': task-1117511, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.102663} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 588.255908] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8d3fe75-9fbe-4167-b227-285d91eb9660 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 588.255908] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a8d3fe75-9fbe-4167-b227-285d91eb9660 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] [instance: 5b86b63a-f6c1-4983-a3cc-3cde7a10d17b] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 588.255908] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a8d3fe75-9fbe-4167-b227-285d91eb9660 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] [instance: 5b86b63a-f6c1-4983-a3cc-3cde7a10d17b] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 588.255908] env[63531]: INFO nova.compute.manager [None req-a8d3fe75-9fbe-4167-b227-285d91eb9660 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] [instance: 5b86b63a-f6c1-4983-a3cc-3cde7a10d17b] Took 1.09 seconds to destroy the instance on the hypervisor. [ 588.255908] env[63531]: DEBUG oslo.service.loopingcall [None req-a8d3fe75-9fbe-4167-b227-285d91eb9660 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 588.256127] env[63531]: DEBUG nova.compute.manager [-] [instance: 5b86b63a-f6c1-4983-a3cc-3cde7a10d17b] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 588.256127] env[63531]: DEBUG nova.network.neutron [-] [instance: 5b86b63a-f6c1-4983-a3cc-3cde7a10d17b] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 588.279622] env[63531]: DEBUG nova.network.neutron [-] [instance: 5b86b63a-f6c1-4983-a3cc-3cde7a10d17b] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 588.333807] env[63531]: DEBUG oslo_concurrency.lockutils [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] Acquiring lock "c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 588.334075] env[63531]: DEBUG oslo_concurrency.lockutils [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] Lock "c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 588.525943] env[63531]: DEBUG oslo_vmware.api [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Task: {'id': task-1117512, 'name': ReconfigVM_Task, 'duration_secs': 0.294336} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 588.526270] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Reconfigured VM instance instance-00000003 to attach disk [datastore1] 4410b013-868a-41e7-9dfb-584fb2d9d142/4410b013-868a-41e7-9dfb-584fb2d9d142.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 588.526860] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-eb1d4248-ba93-4f1d-87a6-c5dee23d72d6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.536044] env[63531]: DEBUG oslo_vmware.api [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Waiting for the task: (returnval){ [ 588.536044] env[63531]: value = "task-1117513" [ 588.536044] env[63531]: _type = "Task" [ 588.536044] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 588.555346] env[63531]: DEBUG oslo_vmware.api [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Task: {'id': task-1117513, 'name': Rename_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 588.708133] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 316c12f7-bc05-4f70-98b4-579d4a7b69d0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 588.708133] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=63531) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 588.708133] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=63531) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 588.784759] env[63531]: DEBUG nova.network.neutron [-] [instance: 5b86b63a-f6c1-4983-a3cc-3cde7a10d17b] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 589.051509] env[63531]: DEBUG oslo_vmware.api [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Task: {'id': task-1117513, 'name': Rename_Task, 'duration_secs': 0.145599} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 589.051509] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 589.051509] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f77da469-984a-45d1-8e33-ebb27c6295d5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.058498] env[63531]: DEBUG oslo_vmware.api [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Waiting for the task: (returnval){ [ 589.058498] env[63531]: value = "task-1117514" [ 589.058498] env[63531]: _type = "Task" [ 589.058498] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 589.076811] env[63531]: DEBUG oslo_vmware.api [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Task: {'id': task-1117514, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 589.108133] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19d3b6fd-2fc6-435e-a59a-bfdc1de45c85 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.119621] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24ccd45b-a2e8-4b72-ae02-3460860f1ce3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.171075] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ebe1a11-addd-4da1-a3f3-aa8b99d7faae {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.178094] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3bd7247-0491-4dab-9ea9-a009bd1a3364 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.197586] env[63531]: DEBUG nova.compute.provider_tree [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 589.290281] env[63531]: INFO nova.compute.manager [-] [instance: 5b86b63a-f6c1-4983-a3cc-3cde7a10d17b] Took 1.04 seconds to deallocate network for instance. [ 589.573249] env[63531]: DEBUG oslo_vmware.api [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Task: {'id': task-1117514, 'name': PowerOnVM_Task} progress is 78%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 589.700381] env[63531]: DEBUG nova.scheduler.client.report [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 589.799619] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a8d3fe75-9fbe-4167-b227-285d91eb9660 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 590.074840] env[63531]: DEBUG oslo_vmware.api [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Task: {'id': task-1117514, 'name': PowerOnVM_Task, 'duration_secs': 0.608501} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 590.074840] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 590.078756] env[63531]: DEBUG nova.compute.manager [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 590.079237] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d9678a5-594f-43bf-913e-ea7f7e1a9818 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.205677] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63531) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 590.205938] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 9.651s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 590.206253] env[63531]: DEBUG oslo_concurrency.lockutils [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.775s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 590.207927] env[63531]: INFO nova.compute.claims [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 590.609443] env[63531]: DEBUG oslo_concurrency.lockutils [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 590.798649] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] Acquiring lock "5521c3ab-d017-473d-b8ce-d37c8b2bcc8f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 590.798939] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] Lock "5521c3ab-d017-473d-b8ce-d37c8b2bcc8f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 591.572904] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09a23608-a527-4e11-a6d7-5f4b2ba00432 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.581739] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-755d8e4c-7637-4b1b-a81a-9ccbc296ff68 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.618027] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a14f9cb3-99d2-4f2b-82f5-823dd875ae96 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.626903] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be482859-e465-41f1-949f-f6593d19c942 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.647148] env[63531]: DEBUG nova.compute.provider_tree [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 592.155350] env[63531]: DEBUG nova.scheduler.client.report [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 592.659142] env[63531]: DEBUG oslo_concurrency.lockutils [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.452s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 592.661318] env[63531]: DEBUG nova.compute.manager [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 592.666360] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.979s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 592.669140] env[63531]: INFO nova.compute.claims [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] [instance: d941a593-7047-436a-b95d-bcb4850829f5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 592.934151] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] Acquiring lock "9a4f813b-5aac-4458-8fc3-af33b4b9636c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 592.934242] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] Lock "9a4f813b-5aac-4458-8fc3-af33b4b9636c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 593.180416] env[63531]: DEBUG nova.compute.utils [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 593.185057] env[63531]: DEBUG nova.compute.manager [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 593.188021] env[63531]: DEBUG nova.network.neutron [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 593.230368] env[63531]: DEBUG nova.policy [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '07142a4db1924c1484eb60ee55aec06c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cf90f907de64448ca9c01029f694ed6c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 593.544222] env[63531]: DEBUG nova.network.neutron [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] Successfully created port: f474eb4d-6d17-448a-8665-0b00d2223b52 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 593.688250] env[63531]: DEBUG nova.compute.manager [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 594.008946] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3782c09a-2989-4864-bda4-f4769d093c3b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.019915] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80d3937e-d5c0-4e71-80d9-ab6653b40722 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.059907] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bde2b9a3-8564-440b-b6ca-4f2de9f9ea2b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.071872] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af8f9b53-e790-42c9-8edf-338900ef3aa2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.089466] env[63531]: DEBUG nova.compute.provider_tree [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 594.200510] env[63531]: DEBUG oslo_concurrency.lockutils [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Acquiring lock "a4dfae30-58f9-4e0b-88b7-c2969d267df6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 594.200779] env[63531]: DEBUG oslo_concurrency.lockutils [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Lock "a4dfae30-58f9-4e0b-88b7-c2969d267df6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 594.595023] env[63531]: DEBUG nova.scheduler.client.report [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 594.703409] env[63531]: DEBUG nova.compute.manager [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 594.741430] env[63531]: DEBUG nova.virt.hardware [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 594.741695] env[63531]: DEBUG nova.virt.hardware [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 594.742010] env[63531]: DEBUG nova.virt.hardware [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 594.742092] env[63531]: DEBUG nova.virt.hardware [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 594.742344] env[63531]: DEBUG nova.virt.hardware [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 594.742470] env[63531]: DEBUG nova.virt.hardware [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 594.742765] env[63531]: DEBUG nova.virt.hardware [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 594.743086] env[63531]: DEBUG nova.virt.hardware [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 594.743158] env[63531]: DEBUG nova.virt.hardware [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 594.743387] env[63531]: DEBUG nova.virt.hardware [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 594.743551] env[63531]: DEBUG nova.virt.hardware [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 594.744598] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62ae0a2c-0879-4dae-877d-46f59a8796b2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.755619] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dda1e82b-0c5a-4e79-a300-deea55a1aa34 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.105158] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.434s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 595.105158] env[63531]: DEBUG nova.compute.manager [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] [instance: d941a593-7047-436a-b95d-bcb4850829f5] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 595.105158] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 23.414s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 595.258465] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Acquiring lock "2a329146-5e0d-4a1f-ba65-3efda2f5e1e0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 595.258691] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Lock "2a329146-5e0d-4a1f-ba65-3efda2f5e1e0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 595.411465] env[63531]: ERROR nova.compute.manager [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f474eb4d-6d17-448a-8665-0b00d2223b52, please check neutron logs for more information. [ 595.411465] env[63531]: ERROR nova.compute.manager Traceback (most recent call last): [ 595.411465] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 595.411465] env[63531]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 595.411465] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 595.411465] env[63531]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 595.411465] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 595.411465] env[63531]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 595.411465] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 595.411465] env[63531]: ERROR nova.compute.manager self.force_reraise() [ 595.411465] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 595.411465] env[63531]: ERROR nova.compute.manager raise self.value [ 595.411465] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 595.411465] env[63531]: ERROR nova.compute.manager updated_port = self._update_port( [ 595.411465] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 595.411465] env[63531]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 595.413105] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 595.413105] env[63531]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 595.413105] env[63531]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f474eb4d-6d17-448a-8665-0b00d2223b52, please check neutron logs for more information. [ 595.413105] env[63531]: ERROR nova.compute.manager [ 595.413105] env[63531]: Traceback (most recent call last): [ 595.413105] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 595.413105] env[63531]: listener.cb(fileno) [ 595.413105] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 595.413105] env[63531]: result = function(*args, **kwargs) [ 595.413105] env[63531]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 595.413105] env[63531]: return func(*args, **kwargs) [ 595.413105] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 595.413105] env[63531]: raise e [ 595.413105] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 595.413105] env[63531]: nwinfo = self.network_api.allocate_for_instance( [ 595.413105] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 595.413105] env[63531]: created_port_ids = self._update_ports_for_instance( [ 595.413105] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 595.413105] env[63531]: with excutils.save_and_reraise_exception(): [ 595.413105] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 595.413105] env[63531]: self.force_reraise() [ 595.413105] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 595.413105] env[63531]: raise self.value [ 595.413105] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 595.413105] env[63531]: updated_port = self._update_port( [ 595.413105] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 595.413105] env[63531]: _ensure_no_port_binding_failure(port) [ 595.413105] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 595.413105] env[63531]: raise exception.PortBindingFailed(port_id=port['id']) [ 595.413964] env[63531]: nova.exception.PortBindingFailed: Binding failed for port f474eb4d-6d17-448a-8665-0b00d2223b52, please check neutron logs for more information. [ 595.413964] env[63531]: Removing descriptor: 17 [ 595.413964] env[63531]: ERROR nova.compute.manager [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f474eb4d-6d17-448a-8665-0b00d2223b52, please check neutron logs for more information. [ 595.413964] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] Traceback (most recent call last): [ 595.413964] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 595.413964] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] yield resources [ 595.413964] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 595.413964] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] self.driver.spawn(context, instance, image_meta, [ 595.413964] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 595.413964] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 595.413964] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 595.413964] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] vm_ref = self.build_virtual_machine(instance, [ 595.414330] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 595.414330] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] vif_infos = vmwarevif.get_vif_info(self._session, [ 595.414330] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 595.414330] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] for vif in network_info: [ 595.414330] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 595.414330] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] return self._sync_wrapper(fn, *args, **kwargs) [ 595.414330] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 595.414330] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] self.wait() [ 595.414330] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 595.414330] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] self[:] = self._gt.wait() [ 595.414330] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 595.414330] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] return self._exit_event.wait() [ 595.414330] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 595.414709] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] result = hub.switch() [ 595.414709] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 595.414709] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] return self.greenlet.switch() [ 595.414709] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 595.414709] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] result = function(*args, **kwargs) [ 595.414709] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 595.414709] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] return func(*args, **kwargs) [ 595.414709] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 595.414709] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] raise e [ 595.414709] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 595.414709] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] nwinfo = self.network_api.allocate_for_instance( [ 595.414709] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 595.414709] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] created_port_ids = self._update_ports_for_instance( [ 595.415113] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 595.415113] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] with excutils.save_and_reraise_exception(): [ 595.415113] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 595.415113] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] self.force_reraise() [ 595.415113] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 595.415113] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] raise self.value [ 595.415113] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 595.415113] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] updated_port = self._update_port( [ 595.415113] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 595.415113] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] _ensure_no_port_binding_failure(port) [ 595.415113] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 595.415113] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] raise exception.PortBindingFailed(port_id=port['id']) [ 595.415467] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] nova.exception.PortBindingFailed: Binding failed for port f474eb4d-6d17-448a-8665-0b00d2223b52, please check neutron logs for more information. [ 595.415467] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] [ 595.415467] env[63531]: INFO nova.compute.manager [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] Terminating instance [ 595.418497] env[63531]: DEBUG oslo_concurrency.lockutils [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Acquiring lock "refresh_cache-9fa373b3-8e15-4d26-82be-7dda51bc6cf6" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 595.420255] env[63531]: DEBUG oslo_concurrency.lockutils [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Acquired lock "refresh_cache-9fa373b3-8e15-4d26-82be-7dda51bc6cf6" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 595.420255] env[63531]: DEBUG nova.network.neutron [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 595.553887] env[63531]: DEBUG oslo_concurrency.lockutils [None req-da9f7a66-33b6-42ee-90cf-f8ce53a3c4c8 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Acquiring lock "4410b013-868a-41e7-9dfb-584fb2d9d142" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 595.553887] env[63531]: DEBUG oslo_concurrency.lockutils [None req-da9f7a66-33b6-42ee-90cf-f8ce53a3c4c8 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Lock "4410b013-868a-41e7-9dfb-584fb2d9d142" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 595.553887] env[63531]: DEBUG oslo_concurrency.lockutils [None req-da9f7a66-33b6-42ee-90cf-f8ce53a3c4c8 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Acquiring lock "4410b013-868a-41e7-9dfb-584fb2d9d142-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 595.553887] env[63531]: DEBUG oslo_concurrency.lockutils [None req-da9f7a66-33b6-42ee-90cf-f8ce53a3c4c8 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Lock "4410b013-868a-41e7-9dfb-584fb2d9d142-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 595.554225] env[63531]: DEBUG oslo_concurrency.lockutils [None req-da9f7a66-33b6-42ee-90cf-f8ce53a3c4c8 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Lock "4410b013-868a-41e7-9dfb-584fb2d9d142-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 595.555601] env[63531]: INFO nova.compute.manager [None req-da9f7a66-33b6-42ee-90cf-f8ce53a3c4c8 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Terminating instance [ 595.559605] env[63531]: DEBUG oslo_concurrency.lockutils [None req-da9f7a66-33b6-42ee-90cf-f8ce53a3c4c8 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Acquiring lock "refresh_cache-4410b013-868a-41e7-9dfb-584fb2d9d142" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 595.559605] env[63531]: DEBUG oslo_concurrency.lockutils [None req-da9f7a66-33b6-42ee-90cf-f8ce53a3c4c8 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Acquired lock "refresh_cache-4410b013-868a-41e7-9dfb-584fb2d9d142" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 595.559605] env[63531]: DEBUG nova.network.neutron [None req-da9f7a66-33b6-42ee-90cf-f8ce53a3c4c8 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 595.613057] env[63531]: DEBUG nova.compute.utils [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 595.613460] env[63531]: DEBUG nova.compute.manager [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] [instance: d941a593-7047-436a-b95d-bcb4850829f5] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 595.613793] env[63531]: DEBUG nova.network.neutron [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] [instance: d941a593-7047-436a-b95d-bcb4850829f5] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 595.722791] env[63531]: DEBUG nova.policy [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c23f81bb55f042148cf4dead270858f8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '16739298b53f4c4aa1ce21ef1c70134d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 595.995198] env[63531]: DEBUG nova.network.neutron [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 596.030021] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-010b4e15-1509-428a-b753-d38b27b125ad {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.040234] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e0f075f-8e92-41f6-a98c-2a9679b4a392 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.080468] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0614af01-8907-4153-a6b3-5468b8a0a494 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.088673] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-254a5c25-bbe6-4c74-8ec1-e6a798a6cdee {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.108510] env[63531]: DEBUG nova.compute.provider_tree [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 596.110119] env[63531]: DEBUG nova.network.neutron [None req-da9f7a66-33b6-42ee-90cf-f8ce53a3c4c8 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 596.124412] env[63531]: DEBUG nova.compute.manager [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] [instance: d941a593-7047-436a-b95d-bcb4850829f5] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 596.246293] env[63531]: DEBUG nova.network.neutron [None req-da9f7a66-33b6-42ee-90cf-f8ce53a3c4c8 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 596.326873] env[63531]: DEBUG nova.network.neutron [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 596.381224] env[63531]: DEBUG nova.compute.manager [req-06b98b8e-3306-4447-bab4-18208ef82a52 req-9545f7f0-c3ad-4d78-9df1-78fbb5426560 service nova] [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] Received event network-changed-f474eb4d-6d17-448a-8665-0b00d2223b52 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 596.381478] env[63531]: DEBUG nova.compute.manager [req-06b98b8e-3306-4447-bab4-18208ef82a52 req-9545f7f0-c3ad-4d78-9df1-78fbb5426560 service nova] [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] Refreshing instance network info cache due to event network-changed-f474eb4d-6d17-448a-8665-0b00d2223b52. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 596.381610] env[63531]: DEBUG oslo_concurrency.lockutils [req-06b98b8e-3306-4447-bab4-18208ef82a52 req-9545f7f0-c3ad-4d78-9df1-78fbb5426560 service nova] Acquiring lock "refresh_cache-9fa373b3-8e15-4d26-82be-7dda51bc6cf6" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 596.547835] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Acquiring lock "3f3113dd-863f-442d-a214-28dde377d227" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 596.548117] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Lock "3f3113dd-863f-442d-a214-28dde377d227" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 596.613598] env[63531]: DEBUG nova.scheduler.client.report [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 596.712106] env[63531]: DEBUG nova.network.neutron [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] [instance: d941a593-7047-436a-b95d-bcb4850829f5] Successfully created port: 38892f48-025c-4a63-b5ca-422b27c6177d {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 596.749610] env[63531]: DEBUG oslo_concurrency.lockutils [None req-da9f7a66-33b6-42ee-90cf-f8ce53a3c4c8 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Releasing lock "refresh_cache-4410b013-868a-41e7-9dfb-584fb2d9d142" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 596.750028] env[63531]: DEBUG nova.compute.manager [None req-da9f7a66-33b6-42ee-90cf-f8ce53a3c4c8 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 596.750191] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-da9f7a66-33b6-42ee-90cf-f8ce53a3c4c8 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 596.751192] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0eb141f-6657-4615-aaa3-b294aad18b07 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.764064] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-da9f7a66-33b6-42ee-90cf-f8ce53a3c4c8 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 596.764064] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d616a9d4-2537-4b40-9958-f14baed89ad8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.773074] env[63531]: DEBUG oslo_vmware.api [None req-da9f7a66-33b6-42ee-90cf-f8ce53a3c4c8 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Waiting for the task: (returnval){ [ 596.773074] env[63531]: value = "task-1117518" [ 596.773074] env[63531]: _type = "Task" [ 596.773074] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 596.785438] env[63531]: DEBUG oslo_vmware.api [None req-da9f7a66-33b6-42ee-90cf-f8ce53a3c4c8 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': task-1117518, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 596.836752] env[63531]: DEBUG oslo_concurrency.lockutils [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Releasing lock "refresh_cache-9fa373b3-8e15-4d26-82be-7dda51bc6cf6" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 596.836752] env[63531]: DEBUG nova.compute.manager [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 596.836752] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 596.836752] env[63531]: DEBUG oslo_concurrency.lockutils [req-06b98b8e-3306-4447-bab4-18208ef82a52 req-9545f7f0-c3ad-4d78-9df1-78fbb5426560 service nova] Acquired lock "refresh_cache-9fa373b3-8e15-4d26-82be-7dda51bc6cf6" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 596.836752] env[63531]: DEBUG nova.network.neutron [req-06b98b8e-3306-4447-bab4-18208ef82a52 req-9545f7f0-c3ad-4d78-9df1-78fbb5426560 service nova] [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] Refreshing network info cache for port f474eb4d-6d17-448a-8665-0b00d2223b52 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 596.838524] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-589c0c41-b0e6-4fe8-b1c1-35e605c6a27d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.849891] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3265fa76-53d1-40cb-b048-4c2f3ae7a75c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.878545] env[63531]: WARNING nova.virt.vmwareapi.vmops [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9fa373b3-8e15-4d26-82be-7dda51bc6cf6 could not be found. [ 596.883160] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 596.883160] env[63531]: INFO nova.compute.manager [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] Took 0.05 seconds to destroy the instance on the hypervisor. [ 596.883160] env[63531]: DEBUG oslo.service.loopingcall [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 596.883160] env[63531]: DEBUG nova.compute.manager [-] [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 596.883160] env[63531]: DEBUG nova.network.neutron [-] [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 596.920905] env[63531]: DEBUG nova.network.neutron [-] [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 597.120310] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.015s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 597.120310] env[63531]: ERROR nova.compute.manager [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 670e808d-7316-4724-9751-f9f2a9c1c9ac, please check neutron logs for more information. [ 597.120310] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] Traceback (most recent call last): [ 597.120310] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 597.120310] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] self.driver.spawn(context, instance, image_meta, [ 597.120310] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 597.120310] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] self._vmops.spawn(context, instance, image_meta, injected_files, [ 597.120310] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 597.120310] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] vm_ref = self.build_virtual_machine(instance, [ 597.120682] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 597.120682] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] vif_infos = vmwarevif.get_vif_info(self._session, [ 597.120682] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 597.120682] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] for vif in network_info: [ 597.120682] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 597.120682] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] return self._sync_wrapper(fn, *args, **kwargs) [ 597.120682] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 597.120682] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] self.wait() [ 597.120682] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 597.120682] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] self[:] = self._gt.wait() [ 597.120682] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 597.120682] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] return self._exit_event.wait() [ 597.120682] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 597.121053] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] result = hub.switch() [ 597.121053] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 597.121053] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] return self.greenlet.switch() [ 597.121053] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 597.121053] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] result = function(*args, **kwargs) [ 597.121053] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 597.121053] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] return func(*args, **kwargs) [ 597.121053] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 597.121053] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] raise e [ 597.121053] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 597.121053] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] nwinfo = self.network_api.allocate_for_instance( [ 597.121053] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 597.121053] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] created_port_ids = self._update_ports_for_instance( [ 597.121474] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 597.121474] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] with excutils.save_and_reraise_exception(): [ 597.121474] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 597.121474] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] self.force_reraise() [ 597.121474] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 597.121474] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] raise self.value [ 597.121474] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 597.121474] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] updated_port = self._update_port( [ 597.121474] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 597.121474] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] _ensure_no_port_binding_failure(port) [ 597.121474] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 597.121474] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] raise exception.PortBindingFailed(port_id=port['id']) [ 597.121908] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] nova.exception.PortBindingFailed: Binding failed for port 670e808d-7316-4724-9751-f9f2a9c1c9ac, please check neutron logs for more information. [ 597.121908] env[63531]: ERROR nova.compute.manager [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] [ 597.123586] env[63531]: DEBUG nova.compute.utils [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] Binding failed for port 670e808d-7316-4724-9751-f9f2a9c1c9ac, please check neutron logs for more information. {{(pid=63531) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 597.124149] env[63531]: DEBUG oslo_concurrency.lockutils [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.476s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 597.125656] env[63531]: INFO nova.compute.claims [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 597.130159] env[63531]: DEBUG nova.compute.manager [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] Build of instance 6a380967-bca6-47d2-b052-988e6c1004ee was re-scheduled: Binding failed for port 670e808d-7316-4724-9751-f9f2a9c1c9ac, please check neutron logs for more information. {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 597.130552] env[63531]: DEBUG nova.compute.manager [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] Unplugging VIFs for instance {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 597.130747] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Acquiring lock "refresh_cache-6a380967-bca6-47d2-b052-988e6c1004ee" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 597.130859] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Acquired lock "refresh_cache-6a380967-bca6-47d2-b052-988e6c1004ee" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 597.131034] env[63531]: DEBUG nova.network.neutron [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 597.144209] env[63531]: DEBUG nova.compute.manager [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] [instance: d941a593-7047-436a-b95d-bcb4850829f5] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 597.185864] env[63531]: DEBUG nova.virt.hardware [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 597.186092] env[63531]: DEBUG nova.virt.hardware [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 597.186251] env[63531]: DEBUG nova.virt.hardware [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 597.186433] env[63531]: DEBUG nova.virt.hardware [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 597.186576] env[63531]: DEBUG nova.virt.hardware [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 597.186731] env[63531]: DEBUG nova.virt.hardware [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 597.186948] env[63531]: DEBUG nova.virt.hardware [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 597.188416] env[63531]: DEBUG nova.virt.hardware [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 597.188976] env[63531]: DEBUG nova.virt.hardware [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 597.188976] env[63531]: DEBUG nova.virt.hardware [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 597.189091] env[63531]: DEBUG nova.virt.hardware [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 597.191212] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcb24494-d9cd-444f-955b-9d248f5514cd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.204253] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5a1ebfc-8c73-48c8-ad27-ae322a65dcf6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.285644] env[63531]: DEBUG oslo_vmware.api [None req-da9f7a66-33b6-42ee-90cf-f8ce53a3c4c8 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': task-1117518, 'name': PowerOffVM_Task, 'duration_secs': 0.156785} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 597.285903] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-da9f7a66-33b6-42ee-90cf-f8ce53a3c4c8 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 597.286075] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-da9f7a66-33b6-42ee-90cf-f8ce53a3c4c8 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 597.286939] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0bf13cca-12f8-494f-b5e2-78a1b0eb319e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.356788] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-da9f7a66-33b6-42ee-90cf-f8ce53a3c4c8 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 597.357017] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-da9f7a66-33b6-42ee-90cf-f8ce53a3c4c8 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 597.358555] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-da9f7a66-33b6-42ee-90cf-f8ce53a3c4c8 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Deleting the datastore file [datastore1] 4410b013-868a-41e7-9dfb-584fb2d9d142 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 597.358764] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0d7cce1f-9964-4eda-887f-a154a329bc6e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.370098] env[63531]: DEBUG oslo_vmware.api [None req-da9f7a66-33b6-42ee-90cf-f8ce53a3c4c8 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Waiting for the task: (returnval){ [ 597.370098] env[63531]: value = "task-1117520" [ 597.370098] env[63531]: _type = "Task" [ 597.370098] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 597.370098] env[63531]: DEBUG nova.network.neutron [req-06b98b8e-3306-4447-bab4-18208ef82a52 req-9545f7f0-c3ad-4d78-9df1-78fbb5426560 service nova] [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 597.383216] env[63531]: DEBUG oslo_vmware.api [None req-da9f7a66-33b6-42ee-90cf-f8ce53a3c4c8 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': task-1117520, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 597.426710] env[63531]: DEBUG nova.network.neutron [-] [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 597.512205] env[63531]: DEBUG nova.network.neutron [req-06b98b8e-3306-4447-bab4-18208ef82a52 req-9545f7f0-c3ad-4d78-9df1-78fbb5426560 service nova] [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 597.671970] env[63531]: DEBUG nova.network.neutron [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 597.750047] env[63531]: DEBUG nova.network.neutron [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 597.888254] env[63531]: DEBUG oslo_vmware.api [None req-da9f7a66-33b6-42ee-90cf-f8ce53a3c4c8 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Task: {'id': task-1117520, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.152116} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 597.888691] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-da9f7a66-33b6-42ee-90cf-f8ce53a3c4c8 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 597.889014] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-da9f7a66-33b6-42ee-90cf-f8ce53a3c4c8 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 597.889085] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-da9f7a66-33b6-42ee-90cf-f8ce53a3c4c8 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 597.889261] env[63531]: INFO nova.compute.manager [None req-da9f7a66-33b6-42ee-90cf-f8ce53a3c4c8 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Took 1.14 seconds to destroy the instance on the hypervisor. [ 597.889502] env[63531]: DEBUG oslo.service.loopingcall [None req-da9f7a66-33b6-42ee-90cf-f8ce53a3c4c8 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 597.889687] env[63531]: DEBUG nova.compute.manager [-] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 597.889875] env[63531]: DEBUG nova.network.neutron [-] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 597.907386] env[63531]: DEBUG nova.network.neutron [-] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 597.931121] env[63531]: INFO nova.compute.manager [-] [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] Took 1.05 seconds to deallocate network for instance. [ 597.934184] env[63531]: DEBUG nova.compute.claims [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] Aborting claim: {{(pid=63531) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 597.934278] env[63531]: DEBUG oslo_concurrency.lockutils [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 598.016166] env[63531]: DEBUG oslo_concurrency.lockutils [req-06b98b8e-3306-4447-bab4-18208ef82a52 req-9545f7f0-c3ad-4d78-9df1-78fbb5426560 service nova] Releasing lock "refresh_cache-9fa373b3-8e15-4d26-82be-7dda51bc6cf6" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 598.097612] env[63531]: ERROR nova.compute.manager [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 38892f48-025c-4a63-b5ca-422b27c6177d, please check neutron logs for more information. [ 598.097612] env[63531]: ERROR nova.compute.manager Traceback (most recent call last): [ 598.097612] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 598.097612] env[63531]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 598.097612] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 598.097612] env[63531]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 598.097612] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 598.097612] env[63531]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 598.097612] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 598.097612] env[63531]: ERROR nova.compute.manager self.force_reraise() [ 598.097612] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 598.097612] env[63531]: ERROR nova.compute.manager raise self.value [ 598.097612] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 598.097612] env[63531]: ERROR nova.compute.manager updated_port = self._update_port( [ 598.097612] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 598.097612] env[63531]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 598.098112] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 598.098112] env[63531]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 598.098112] env[63531]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 38892f48-025c-4a63-b5ca-422b27c6177d, please check neutron logs for more information. [ 598.098112] env[63531]: ERROR nova.compute.manager [ 598.098112] env[63531]: Traceback (most recent call last): [ 598.098112] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 598.098112] env[63531]: listener.cb(fileno) [ 598.098112] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 598.098112] env[63531]: result = function(*args, **kwargs) [ 598.098112] env[63531]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 598.098112] env[63531]: return func(*args, **kwargs) [ 598.098112] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 598.098112] env[63531]: raise e [ 598.098112] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 598.098112] env[63531]: nwinfo = self.network_api.allocate_for_instance( [ 598.098112] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 598.098112] env[63531]: created_port_ids = self._update_ports_for_instance( [ 598.098112] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 598.098112] env[63531]: with excutils.save_and_reraise_exception(): [ 598.098112] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 598.098112] env[63531]: self.force_reraise() [ 598.098112] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 598.098112] env[63531]: raise self.value [ 598.098112] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 598.098112] env[63531]: updated_port = self._update_port( [ 598.098112] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 598.098112] env[63531]: _ensure_no_port_binding_failure(port) [ 598.098112] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 598.098112] env[63531]: raise exception.PortBindingFailed(port_id=port['id']) [ 598.098930] env[63531]: nova.exception.PortBindingFailed: Binding failed for port 38892f48-025c-4a63-b5ca-422b27c6177d, please check neutron logs for more information. [ 598.098930] env[63531]: Removing descriptor: 17 [ 598.098930] env[63531]: ERROR nova.compute.manager [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] [instance: d941a593-7047-436a-b95d-bcb4850829f5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 38892f48-025c-4a63-b5ca-422b27c6177d, please check neutron logs for more information. [ 598.098930] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] Traceback (most recent call last): [ 598.098930] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 598.098930] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] yield resources [ 598.098930] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 598.098930] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] self.driver.spawn(context, instance, image_meta, [ 598.098930] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 598.098930] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 598.098930] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 598.098930] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] vm_ref = self.build_virtual_machine(instance, [ 598.099283] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 598.099283] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] vif_infos = vmwarevif.get_vif_info(self._session, [ 598.099283] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 598.099283] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] for vif in network_info: [ 598.099283] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 598.099283] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] return self._sync_wrapper(fn, *args, **kwargs) [ 598.099283] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 598.099283] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] self.wait() [ 598.099283] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 598.099283] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] self[:] = self._gt.wait() [ 598.099283] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 598.099283] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] return self._exit_event.wait() [ 598.099283] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 598.099637] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] result = hub.switch() [ 598.099637] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 598.099637] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] return self.greenlet.switch() [ 598.099637] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 598.099637] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] result = function(*args, **kwargs) [ 598.099637] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 598.099637] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] return func(*args, **kwargs) [ 598.099637] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 598.099637] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] raise e [ 598.099637] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 598.099637] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] nwinfo = self.network_api.allocate_for_instance( [ 598.099637] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 598.099637] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] created_port_ids = self._update_ports_for_instance( [ 598.099994] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 598.099994] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] with excutils.save_and_reraise_exception(): [ 598.099994] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 598.099994] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] self.force_reraise() [ 598.099994] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 598.099994] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] raise self.value [ 598.099994] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 598.099994] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] updated_port = self._update_port( [ 598.099994] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 598.099994] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] _ensure_no_port_binding_failure(port) [ 598.099994] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 598.099994] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] raise exception.PortBindingFailed(port_id=port['id']) [ 598.100347] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] nova.exception.PortBindingFailed: Binding failed for port 38892f48-025c-4a63-b5ca-422b27c6177d, please check neutron logs for more information. [ 598.100347] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] [ 598.100347] env[63531]: INFO nova.compute.manager [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] [instance: d941a593-7047-436a-b95d-bcb4850829f5] Terminating instance [ 598.101750] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] Acquiring lock "refresh_cache-d941a593-7047-436a-b95d-bcb4850829f5" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 598.101750] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] Acquired lock "refresh_cache-d941a593-7047-436a-b95d-bcb4850829f5" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 598.101750] env[63531]: DEBUG nova.network.neutron [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] [instance: d941a593-7047-436a-b95d-bcb4850829f5] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 598.254514] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Releasing lock "refresh_cache-6a380967-bca6-47d2-b052-988e6c1004ee" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 598.254514] env[63531]: DEBUG nova.compute.manager [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 598.254514] env[63531]: DEBUG nova.compute.manager [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 598.254514] env[63531]: DEBUG nova.network.neutron [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 598.259625] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Acquiring lock "ffb34f90-d22f-440b-ba07-75d474c3c300" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 598.259985] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Lock "ffb34f90-d22f-440b-ba07-75d474c3c300" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 598.279207] env[63531]: DEBUG nova.network.neutron [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 598.409263] env[63531]: DEBUG nova.network.neutron [-] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.519650] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1cf2f14-c2b1-4349-9816-dc14a8028d4d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.530189] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0ce8b92-8d0e-480e-94a8-f183a875b1f4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.574388] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-036878b0-9cac-4975-b258-a56ba3017808 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.584030] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bfc3a9f-f2ee-40d5-b20d-7250a712724b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.597275] env[63531]: DEBUG nova.compute.provider_tree [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 598.643511] env[63531]: DEBUG nova.network.neutron [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] [instance: d941a593-7047-436a-b95d-bcb4850829f5] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 598.781526] env[63531]: DEBUG nova.network.neutron [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.817810] env[63531]: DEBUG nova.network.neutron [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] [instance: d941a593-7047-436a-b95d-bcb4850829f5] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.914066] env[63531]: INFO nova.compute.manager [-] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Took 1.02 seconds to deallocate network for instance. [ 599.109550] env[63531]: DEBUG nova.scheduler.client.report [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 599.287523] env[63531]: INFO nova.compute.manager [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 6a380967-bca6-47d2-b052-988e6c1004ee] Took 1.03 seconds to deallocate network for instance. [ 599.320136] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] Releasing lock "refresh_cache-d941a593-7047-436a-b95d-bcb4850829f5" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 599.322034] env[63531]: DEBUG nova.compute.manager [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] [instance: d941a593-7047-436a-b95d-bcb4850829f5] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 599.322034] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] [instance: d941a593-7047-436a-b95d-bcb4850829f5] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 599.322034] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-94560893-e71b-49e5-b800-f8c716c09833 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.332435] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cf1a2ca-01ba-4565-bfe2-7c807606c687 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.355993] env[63531]: WARNING nova.virt.vmwareapi.vmops [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] [instance: d941a593-7047-436a-b95d-bcb4850829f5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d941a593-7047-436a-b95d-bcb4850829f5 could not be found. [ 599.356326] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] [instance: d941a593-7047-436a-b95d-bcb4850829f5] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 599.356516] env[63531]: INFO nova.compute.manager [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] [instance: d941a593-7047-436a-b95d-bcb4850829f5] Took 0.04 seconds to destroy the instance on the hypervisor. [ 599.357021] env[63531]: DEBUG oslo.service.loopingcall [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 599.357279] env[63531]: DEBUG nova.compute.manager [-] [instance: d941a593-7047-436a-b95d-bcb4850829f5] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 599.357369] env[63531]: DEBUG nova.network.neutron [-] [instance: d941a593-7047-436a-b95d-bcb4850829f5] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 599.392050] env[63531]: DEBUG nova.network.neutron [-] [instance: d941a593-7047-436a-b95d-bcb4850829f5] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 599.428997] env[63531]: DEBUG oslo_concurrency.lockutils [None req-da9f7a66-33b6-42ee-90cf-f8ce53a3c4c8 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 599.528160] env[63531]: DEBUG nova.compute.manager [req-7f949360-c8a6-471d-ba07-b2d97649fec4 req-ad01d582-bfff-41e4-89f5-6ce704b79b89 service nova] [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] Received event network-vif-deleted-f474eb4d-6d17-448a-8665-0b00d2223b52 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 599.528160] env[63531]: DEBUG nova.compute.manager [req-7f949360-c8a6-471d-ba07-b2d97649fec4 req-ad01d582-bfff-41e4-89f5-6ce704b79b89 service nova] [instance: d941a593-7047-436a-b95d-bcb4850829f5] Received event network-changed-38892f48-025c-4a63-b5ca-422b27c6177d {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 599.529434] env[63531]: DEBUG nova.compute.manager [req-7f949360-c8a6-471d-ba07-b2d97649fec4 req-ad01d582-bfff-41e4-89f5-6ce704b79b89 service nova] [instance: d941a593-7047-436a-b95d-bcb4850829f5] Refreshing instance network info cache due to event network-changed-38892f48-025c-4a63-b5ca-422b27c6177d. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 599.529779] env[63531]: DEBUG oslo_concurrency.lockutils [req-7f949360-c8a6-471d-ba07-b2d97649fec4 req-ad01d582-bfff-41e4-89f5-6ce704b79b89 service nova] Acquiring lock "refresh_cache-d941a593-7047-436a-b95d-bcb4850829f5" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 599.530084] env[63531]: DEBUG oslo_concurrency.lockutils [req-7f949360-c8a6-471d-ba07-b2d97649fec4 req-ad01d582-bfff-41e4-89f5-6ce704b79b89 service nova] Acquired lock "refresh_cache-d941a593-7047-436a-b95d-bcb4850829f5" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 599.530938] env[63531]: DEBUG nova.network.neutron [req-7f949360-c8a6-471d-ba07-b2d97649fec4 req-ad01d582-bfff-41e4-89f5-6ce704b79b89 service nova] [instance: d941a593-7047-436a-b95d-bcb4850829f5] Refreshing network info cache for port 38892f48-025c-4a63-b5ca-422b27c6177d {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 599.618019] env[63531]: DEBUG oslo_concurrency.lockutils [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.491s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 599.618019] env[63531]: DEBUG nova.compute.manager [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 599.618790] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.878s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 599.620420] env[63531]: INFO nova.compute.claims [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 599.729303] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Acquiring lock "0c948655-362d-4673-b3e6-3f28ec69ea3c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 599.729968] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Lock "0c948655-362d-4673-b3e6-3f28ec69ea3c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 599.895464] env[63531]: DEBUG nova.network.neutron [-] [instance: d941a593-7047-436a-b95d-bcb4850829f5] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 600.061724] env[63531]: DEBUG nova.network.neutron [req-7f949360-c8a6-471d-ba07-b2d97649fec4 req-ad01d582-bfff-41e4-89f5-6ce704b79b89 service nova] [instance: d941a593-7047-436a-b95d-bcb4850829f5] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 600.128624] env[63531]: DEBUG nova.compute.utils [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 600.138674] env[63531]: DEBUG nova.compute.manager [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 600.138863] env[63531]: DEBUG nova.network.neutron [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 600.265895] env[63531]: DEBUG nova.network.neutron [req-7f949360-c8a6-471d-ba07-b2d97649fec4 req-ad01d582-bfff-41e4-89f5-6ce704b79b89 service nova] [instance: d941a593-7047-436a-b95d-bcb4850829f5] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 600.280299] env[63531]: DEBUG nova.policy [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '694d4b45031442b8be59b893518a0a42', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2de66f356529416d8de44e672c8f0aa0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 600.324509] env[63531]: INFO nova.scheduler.client.report [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Deleted allocations for instance 6a380967-bca6-47d2-b052-988e6c1004ee [ 600.398905] env[63531]: INFO nova.compute.manager [-] [instance: d941a593-7047-436a-b95d-bcb4850829f5] Took 1.04 seconds to deallocate network for instance. [ 600.403171] env[63531]: DEBUG nova.compute.claims [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] [instance: d941a593-7047-436a-b95d-bcb4850829f5] Aborting claim: {{(pid=63531) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 600.403364] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 600.639709] env[63531]: DEBUG nova.compute.manager [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 600.756398] env[63531]: DEBUG nova.network.neutron [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] Successfully created port: 2c3c54ee-74be-4a1e-bfa8-a24d3e562f20 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 600.772989] env[63531]: DEBUG oslo_concurrency.lockutils [req-7f949360-c8a6-471d-ba07-b2d97649fec4 req-ad01d582-bfff-41e4-89f5-6ce704b79b89 service nova] Releasing lock "refresh_cache-d941a593-7047-436a-b95d-bcb4850829f5" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 600.772989] env[63531]: DEBUG nova.compute.manager [req-7f949360-c8a6-471d-ba07-b2d97649fec4 req-ad01d582-bfff-41e4-89f5-6ce704b79b89 service nova] [instance: d941a593-7047-436a-b95d-bcb4850829f5] Received event network-vif-deleted-38892f48-025c-4a63-b5ca-422b27c6177d {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 600.841892] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ad25429f-a494-491e-a458-815061e31f16 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Lock "6a380967-bca6-47d2-b052-988e6c1004ee" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.537s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 601.052228] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8763d47-0a8f-4ed5-bac2-85bba357540c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.060988] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3ca4638-e031-4030-b5ae-b49ff39b0bd8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.091219] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d67b60f-b670-4d1b-b3a7-b51f03abd4cc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.098844] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11e377fc-21f2-4ce1-bfe0-e4bddd153142 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.115032] env[63531]: DEBUG nova.compute.provider_tree [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 601.346317] env[63531]: DEBUG nova.compute.manager [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 56872ede-96f0-48ca-b289-86784f37527d] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 601.621133] env[63531]: DEBUG nova.scheduler.client.report [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 601.652575] env[63531]: DEBUG nova.compute.manager [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 601.689811] env[63531]: DEBUG nova.virt.hardware [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 601.690119] env[63531]: DEBUG nova.virt.hardware [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 601.690247] env[63531]: DEBUG nova.virt.hardware [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 601.690446] env[63531]: DEBUG nova.virt.hardware [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 601.690603] env[63531]: DEBUG nova.virt.hardware [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 601.690767] env[63531]: DEBUG nova.virt.hardware [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 601.691061] env[63531]: DEBUG nova.virt.hardware [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 601.691181] env[63531]: DEBUG nova.virt.hardware [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 601.691384] env[63531]: DEBUG nova.virt.hardware [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 601.691551] env[63531]: DEBUG nova.virt.hardware [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 601.691768] env[63531]: DEBUG nova.virt.hardware [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 601.692670] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0a6bcd4-b198-46fd-9fca-3ec0517c8c16 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.702322] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0e683f7-5ec5-45aa-9565-c9173ba4ce99 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.881377] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 602.129989] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.511s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 602.133335] env[63531]: DEBUG nova.compute.manager [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 602.138626] env[63531]: DEBUG oslo_concurrency.lockutils [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.921s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.138626] env[63531]: INFO nova.compute.claims [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 602.637302] env[63531]: DEBUG nova.compute.utils [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 602.651031] env[63531]: DEBUG nova.compute.manager [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 602.651031] env[63531]: DEBUG nova.network.neutron [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 602.785704] env[63531]: DEBUG nova.policy [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b6c9eb201c7843cdaf6dca07fd3299e0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b140f1cecdee423c83b11a7e53c2726a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 602.788150] env[63531]: ERROR nova.compute.manager [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2c3c54ee-74be-4a1e-bfa8-a24d3e562f20, please check neutron logs for more information. [ 602.788150] env[63531]: ERROR nova.compute.manager Traceback (most recent call last): [ 602.788150] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 602.788150] env[63531]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 602.788150] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 602.788150] env[63531]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 602.788150] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 602.788150] env[63531]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 602.788150] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 602.788150] env[63531]: ERROR nova.compute.manager self.force_reraise() [ 602.788150] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 602.788150] env[63531]: ERROR nova.compute.manager raise self.value [ 602.788150] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 602.788150] env[63531]: ERROR nova.compute.manager updated_port = self._update_port( [ 602.788150] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 602.788150] env[63531]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 602.788674] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 602.788674] env[63531]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 602.788674] env[63531]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2c3c54ee-74be-4a1e-bfa8-a24d3e562f20, please check neutron logs for more information. [ 602.788674] env[63531]: ERROR nova.compute.manager [ 602.788674] env[63531]: Traceback (most recent call last): [ 602.788674] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 602.788674] env[63531]: listener.cb(fileno) [ 602.788674] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 602.788674] env[63531]: result = function(*args, **kwargs) [ 602.788674] env[63531]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 602.788674] env[63531]: return func(*args, **kwargs) [ 602.788674] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 602.788674] env[63531]: raise e [ 602.788674] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 602.788674] env[63531]: nwinfo = self.network_api.allocate_for_instance( [ 602.788674] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 602.788674] env[63531]: created_port_ids = self._update_ports_for_instance( [ 602.788674] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 602.788674] env[63531]: with excutils.save_and_reraise_exception(): [ 602.788674] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 602.788674] env[63531]: self.force_reraise() [ 602.788674] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 602.788674] env[63531]: raise self.value [ 602.788674] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 602.788674] env[63531]: updated_port = self._update_port( [ 602.788674] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 602.788674] env[63531]: _ensure_no_port_binding_failure(port) [ 602.788674] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 602.788674] env[63531]: raise exception.PortBindingFailed(port_id=port['id']) [ 602.789629] env[63531]: nova.exception.PortBindingFailed: Binding failed for port 2c3c54ee-74be-4a1e-bfa8-a24d3e562f20, please check neutron logs for more information. [ 602.789629] env[63531]: Removing descriptor: 20 [ 602.789691] env[63531]: ERROR nova.compute.manager [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2c3c54ee-74be-4a1e-bfa8-a24d3e562f20, please check neutron logs for more information. [ 602.789691] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] Traceback (most recent call last): [ 602.789691] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 602.789691] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] yield resources [ 602.789691] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 602.789691] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] self.driver.spawn(context, instance, image_meta, [ 602.789691] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 602.789691] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] self._vmops.spawn(context, instance, image_meta, injected_files, [ 602.789691] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 602.789691] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] vm_ref = self.build_virtual_machine(instance, [ 602.789691] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 602.790303] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] vif_infos = vmwarevif.get_vif_info(self._session, [ 602.790303] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 602.790303] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] for vif in network_info: [ 602.790303] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 602.790303] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] return self._sync_wrapper(fn, *args, **kwargs) [ 602.790303] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 602.790303] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] self.wait() [ 602.790303] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 602.790303] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] self[:] = self._gt.wait() [ 602.790303] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 602.790303] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] return self._exit_event.wait() [ 602.790303] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 602.790303] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] result = hub.switch() [ 602.790712] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 602.790712] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] return self.greenlet.switch() [ 602.790712] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 602.790712] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] result = function(*args, **kwargs) [ 602.790712] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 602.790712] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] return func(*args, **kwargs) [ 602.790712] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 602.790712] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] raise e [ 602.790712] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 602.790712] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] nwinfo = self.network_api.allocate_for_instance( [ 602.790712] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 602.790712] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] created_port_ids = self._update_ports_for_instance( [ 602.790712] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 602.791136] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] with excutils.save_and_reraise_exception(): [ 602.791136] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 602.791136] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] self.force_reraise() [ 602.791136] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 602.791136] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] raise self.value [ 602.791136] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 602.791136] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] updated_port = self._update_port( [ 602.791136] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 602.791136] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] _ensure_no_port_binding_failure(port) [ 602.791136] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 602.791136] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] raise exception.PortBindingFailed(port_id=port['id']) [ 602.791136] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] nova.exception.PortBindingFailed: Binding failed for port 2c3c54ee-74be-4a1e-bfa8-a24d3e562f20, please check neutron logs for more information. [ 602.791136] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] [ 602.791554] env[63531]: INFO nova.compute.manager [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] Terminating instance [ 602.793346] env[63531]: DEBUG oslo_concurrency.lockutils [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Acquiring lock "refresh_cache-2b5e3059-2772-496b-95c7-a083a37ef155" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 602.793346] env[63531]: DEBUG oslo_concurrency.lockutils [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Acquired lock "refresh_cache-2b5e3059-2772-496b-95c7-a083a37ef155" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 602.793346] env[63531]: DEBUG nova.network.neutron [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 603.078457] env[63531]: DEBUG nova.compute.manager [req-0fbeb064-9c44-4876-8fad-38662870fa5d req-8c085ab8-1fa7-4bd1-ac3e-9204c850e009 service nova] [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] Received event network-changed-2c3c54ee-74be-4a1e-bfa8-a24d3e562f20 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 603.078457] env[63531]: DEBUG nova.compute.manager [req-0fbeb064-9c44-4876-8fad-38662870fa5d req-8c085ab8-1fa7-4bd1-ac3e-9204c850e009 service nova] [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] Refreshing instance network info cache due to event network-changed-2c3c54ee-74be-4a1e-bfa8-a24d3e562f20. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 603.078457] env[63531]: DEBUG oslo_concurrency.lockutils [req-0fbeb064-9c44-4876-8fad-38662870fa5d req-8c085ab8-1fa7-4bd1-ac3e-9204c850e009 service nova] Acquiring lock "refresh_cache-2b5e3059-2772-496b-95c7-a083a37ef155" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 603.158173] env[63531]: DEBUG nova.compute.manager [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 603.424340] env[63531]: DEBUG nova.network.neutron [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 603.591115] env[63531]: DEBUG oslo_concurrency.lockutils [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Acquiring lock "7bc63fdd-4b50-486d-9681-ca0baa08f7b3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 603.591350] env[63531]: DEBUG oslo_concurrency.lockutils [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Lock "7bc63fdd-4b50-486d-9681-ca0baa08f7b3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 603.592139] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87acb3c7-5cbf-4b5b-8b8f-05b6b8578570 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.600815] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-772aeeed-7b46-4e60-9278-69800b219b7a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.631961] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf6ff382-815d-4668-bbf8-65b6e7accf53 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.639570] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f40a6450-b8f4-41a2-aab7-64732a3e3d8d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.656813] env[63531]: DEBUG nova.compute.provider_tree [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 603.658875] env[63531]: DEBUG nova.network.neutron [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.723689] env[63531]: DEBUG nova.network.neutron [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] Successfully created port: 3670ef9d-bf2d-4638-a691-74ed5d418a05 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 604.161873] env[63531]: DEBUG nova.scheduler.client.report [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 604.166484] env[63531]: DEBUG oslo_concurrency.lockutils [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Releasing lock "refresh_cache-2b5e3059-2772-496b-95c7-a083a37ef155" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 604.166957] env[63531]: DEBUG nova.compute.manager [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 604.167209] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 604.167762] env[63531]: DEBUG oslo_concurrency.lockutils [req-0fbeb064-9c44-4876-8fad-38662870fa5d req-8c085ab8-1fa7-4bd1-ac3e-9204c850e009 service nova] Acquired lock "refresh_cache-2b5e3059-2772-496b-95c7-a083a37ef155" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 604.167859] env[63531]: DEBUG nova.network.neutron [req-0fbeb064-9c44-4876-8fad-38662870fa5d req-8c085ab8-1fa7-4bd1-ac3e-9204c850e009 service nova] [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] Refreshing network info cache for port 2c3c54ee-74be-4a1e-bfa8-a24d3e562f20 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 604.171797] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-22ee054d-598f-4895-8b6a-3f3b7a90d024 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.176404] env[63531]: DEBUG nova.compute.manager [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 604.190495] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fb7c7b4-90f6-40b9-ba71-c147afa836b9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.223246] env[63531]: DEBUG nova.virt.hardware [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:16:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='903108457',id=25,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-994450239',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 604.223246] env[63531]: DEBUG nova.virt.hardware [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 604.223246] env[63531]: DEBUG nova.virt.hardware [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 604.223690] env[63531]: DEBUG nova.virt.hardware [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 604.223690] env[63531]: DEBUG nova.virt.hardware [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 604.223690] env[63531]: DEBUG nova.virt.hardware [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 604.223690] env[63531]: DEBUG nova.virt.hardware [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 604.223690] env[63531]: DEBUG nova.virt.hardware [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 604.223961] env[63531]: DEBUG nova.virt.hardware [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 604.223961] env[63531]: DEBUG nova.virt.hardware [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 604.223961] env[63531]: DEBUG nova.virt.hardware [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 604.224141] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b2bc948-197c-4473-b5a6-d1607d68a2fa {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.231331] env[63531]: WARNING nova.virt.vmwareapi.vmops [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2b5e3059-2772-496b-95c7-a083a37ef155 could not be found. [ 604.232357] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 604.232357] env[63531]: INFO nova.compute.manager [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] Took 0.06 seconds to destroy the instance on the hypervisor. [ 604.232357] env[63531]: DEBUG oslo.service.loopingcall [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 604.233222] env[63531]: DEBUG nova.compute.manager [-] [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 604.236793] env[63531]: DEBUG nova.network.neutron [-] [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 604.239340] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f1c86e9-f818-4ba9-a6c1-a8c1eccecba5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.285448] env[63531]: DEBUG nova.network.neutron [-] [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 604.669624] env[63531]: DEBUG oslo_concurrency.lockutils [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.533s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 604.670281] env[63531]: DEBUG nova.compute.manager [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 604.673797] env[63531]: DEBUG oslo_concurrency.lockutils [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 25.497s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 604.674132] env[63531]: DEBUG nova.objects.instance [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63531) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 604.715263] env[63531]: DEBUG nova.network.neutron [req-0fbeb064-9c44-4876-8fad-38662870fa5d req-8c085ab8-1fa7-4bd1-ac3e-9204c850e009 service nova] [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 604.789072] env[63531]: DEBUG nova.network.neutron [-] [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.803602] env[63531]: DEBUG oslo_concurrency.lockutils [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Acquiring lock "f0276e34-1de3-482a-82c7-1439d40fd85e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.803843] env[63531]: DEBUG oslo_concurrency.lockutils [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Lock "f0276e34-1de3-482a-82c7-1439d40fd85e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 604.886059] env[63531]: DEBUG nova.network.neutron [req-0fbeb064-9c44-4876-8fad-38662870fa5d req-8c085ab8-1fa7-4bd1-ac3e-9204c850e009 service nova] [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 605.184128] env[63531]: DEBUG nova.compute.utils [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 605.187450] env[63531]: ERROR nova.compute.manager [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3670ef9d-bf2d-4638-a691-74ed5d418a05, please check neutron logs for more information. [ 605.187450] env[63531]: ERROR nova.compute.manager Traceback (most recent call last): [ 605.187450] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 605.187450] env[63531]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 605.187450] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 605.187450] env[63531]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 605.187450] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 605.187450] env[63531]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 605.187450] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 605.187450] env[63531]: ERROR nova.compute.manager self.force_reraise() [ 605.187450] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 605.187450] env[63531]: ERROR nova.compute.manager raise self.value [ 605.187450] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 605.187450] env[63531]: ERROR nova.compute.manager updated_port = self._update_port( [ 605.187450] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 605.187450] env[63531]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 605.187969] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 605.187969] env[63531]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 605.187969] env[63531]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3670ef9d-bf2d-4638-a691-74ed5d418a05, please check neutron logs for more information. [ 605.187969] env[63531]: ERROR nova.compute.manager [ 605.187969] env[63531]: Traceback (most recent call last): [ 605.187969] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 605.187969] env[63531]: listener.cb(fileno) [ 605.187969] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 605.187969] env[63531]: result = function(*args, **kwargs) [ 605.187969] env[63531]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 605.187969] env[63531]: return func(*args, **kwargs) [ 605.187969] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 605.187969] env[63531]: raise e [ 605.187969] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 605.187969] env[63531]: nwinfo = self.network_api.allocate_for_instance( [ 605.187969] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 605.187969] env[63531]: created_port_ids = self._update_ports_for_instance( [ 605.187969] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 605.187969] env[63531]: with excutils.save_and_reraise_exception(): [ 605.187969] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 605.187969] env[63531]: self.force_reraise() [ 605.187969] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 605.187969] env[63531]: raise self.value [ 605.187969] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 605.187969] env[63531]: updated_port = self._update_port( [ 605.187969] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 605.187969] env[63531]: _ensure_no_port_binding_failure(port) [ 605.187969] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 605.187969] env[63531]: raise exception.PortBindingFailed(port_id=port['id']) [ 605.189025] env[63531]: nova.exception.PortBindingFailed: Binding failed for port 3670ef9d-bf2d-4638-a691-74ed5d418a05, please check neutron logs for more information. [ 605.189025] env[63531]: Removing descriptor: 17 [ 605.189025] env[63531]: DEBUG nova.compute.manager [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 605.189025] env[63531]: DEBUG nova.network.neutron [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 605.192890] env[63531]: ERROR nova.compute.manager [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3670ef9d-bf2d-4638-a691-74ed5d418a05, please check neutron logs for more information. [ 605.192890] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] Traceback (most recent call last): [ 605.192890] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 605.192890] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] yield resources [ 605.192890] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 605.192890] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] self.driver.spawn(context, instance, image_meta, [ 605.192890] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 605.192890] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 605.192890] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 605.192890] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] vm_ref = self.build_virtual_machine(instance, [ 605.192890] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 605.194591] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] vif_infos = vmwarevif.get_vif_info(self._session, [ 605.194591] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 605.194591] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] for vif in network_info: [ 605.194591] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 605.194591] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] return self._sync_wrapper(fn, *args, **kwargs) [ 605.194591] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 605.194591] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] self.wait() [ 605.194591] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 605.194591] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] self[:] = self._gt.wait() [ 605.194591] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 605.194591] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] return self._exit_event.wait() [ 605.194591] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 605.194591] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] result = hub.switch() [ 605.195305] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 605.195305] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] return self.greenlet.switch() [ 605.195305] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 605.195305] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] result = function(*args, **kwargs) [ 605.195305] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 605.195305] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] return func(*args, **kwargs) [ 605.195305] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 605.195305] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] raise e [ 605.195305] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 605.195305] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] nwinfo = self.network_api.allocate_for_instance( [ 605.195305] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 605.195305] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] created_port_ids = self._update_ports_for_instance( [ 605.195305] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 605.195889] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] with excutils.save_and_reraise_exception(): [ 605.195889] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 605.195889] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] self.force_reraise() [ 605.195889] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 605.195889] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] raise self.value [ 605.195889] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 605.195889] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] updated_port = self._update_port( [ 605.195889] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 605.195889] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] _ensure_no_port_binding_failure(port) [ 605.195889] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 605.195889] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] raise exception.PortBindingFailed(port_id=port['id']) [ 605.195889] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] nova.exception.PortBindingFailed: Binding failed for port 3670ef9d-bf2d-4638-a691-74ed5d418a05, please check neutron logs for more information. [ 605.195889] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] [ 605.196709] env[63531]: INFO nova.compute.manager [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] Terminating instance [ 605.198701] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] Acquiring lock "refresh_cache-01a1b3e0-4ad9-4350-b59e-499f1b3412a5" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 605.198701] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] Acquired lock "refresh_cache-01a1b3e0-4ad9-4350-b59e-499f1b3412a5" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 605.198701] env[63531]: DEBUG nova.network.neutron [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 605.290976] env[63531]: DEBUG nova.policy [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'adb3093c24784f6688f8e50c576f8e39', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd00af8b9f20f4133a1c644d69c6091d3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 605.296406] env[63531]: INFO nova.compute.manager [-] [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] Took 1.06 seconds to deallocate network for instance. [ 605.297468] env[63531]: DEBUG nova.compute.claims [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] Aborting claim: {{(pid=63531) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 605.297957] env[63531]: DEBUG oslo_concurrency.lockutils [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 605.389793] env[63531]: DEBUG oslo_concurrency.lockutils [req-0fbeb064-9c44-4876-8fad-38662870fa5d req-8c085ab8-1fa7-4bd1-ac3e-9204c850e009 service nova] Releasing lock "refresh_cache-2b5e3059-2772-496b-95c7-a083a37ef155" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 605.689367] env[63531]: DEBUG oslo_concurrency.lockutils [None req-311f0bec-6b63-4724-8871-3f32bc7ddf99 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.015s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 605.691049] env[63531]: DEBUG oslo_concurrency.lockutils [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 24.100s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 605.700647] env[63531]: DEBUG nova.compute.manager [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 605.731941] env[63531]: DEBUG nova.network.neutron [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 605.795788] env[63531]: DEBUG nova.compute.manager [req-400849a6-28be-4837-9eb1-19052e79782d req-42584587-91e7-4e21-88b5-d5cfef1b2142 service nova] [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] Received event network-vif-deleted-2c3c54ee-74be-4a1e-bfa8-a24d3e562f20 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 605.795788] env[63531]: DEBUG nova.compute.manager [req-400849a6-28be-4837-9eb1-19052e79782d req-42584587-91e7-4e21-88b5-d5cfef1b2142 service nova] [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] Received event network-changed-3670ef9d-bf2d-4638-a691-74ed5d418a05 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 605.795788] env[63531]: DEBUG nova.compute.manager [req-400849a6-28be-4837-9eb1-19052e79782d req-42584587-91e7-4e21-88b5-d5cfef1b2142 service nova] [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] Refreshing instance network info cache due to event network-changed-3670ef9d-bf2d-4638-a691-74ed5d418a05. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 605.795788] env[63531]: DEBUG oslo_concurrency.lockutils [req-400849a6-28be-4837-9eb1-19052e79782d req-42584587-91e7-4e21-88b5-d5cfef1b2142 service nova] Acquiring lock "refresh_cache-01a1b3e0-4ad9-4350-b59e-499f1b3412a5" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 605.896322] env[63531]: DEBUG nova.network.neutron [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.111774] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Acquiring lock "c3d597b4-1dc3-48d8-9bee-e73c8929181b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 606.112044] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Lock "c3d597b4-1dc3-48d8-9bee-e73c8929181b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 606.131791] env[63531]: DEBUG nova.network.neutron [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] Successfully created port: c30c9653-38f5-42cc-bda4-656ad1a333fc {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 606.140506] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Acquiring lock "645abb7a-cf73-4fb0-a9a0-49205f060d6f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 606.140750] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Lock "645abb7a-cf73-4fb0-a9a0-49205f060d6f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 606.169941] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Acquiring lock "f39c8807-1e15-41cf-899b-a1fbe0695d58" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 606.170236] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Lock "f39c8807-1e15-41cf-899b-a1fbe0695d58" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 606.402756] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] Releasing lock "refresh_cache-01a1b3e0-4ad9-4350-b59e-499f1b3412a5" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 606.402756] env[63531]: DEBUG nova.compute.manager [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 606.403132] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 606.403211] env[63531]: DEBUG oslo_concurrency.lockutils [req-400849a6-28be-4837-9eb1-19052e79782d req-42584587-91e7-4e21-88b5-d5cfef1b2142 service nova] Acquired lock "refresh_cache-01a1b3e0-4ad9-4350-b59e-499f1b3412a5" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 606.403387] env[63531]: DEBUG nova.network.neutron [req-400849a6-28be-4837-9eb1-19052e79782d req-42584587-91e7-4e21-88b5-d5cfef1b2142 service nova] [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] Refreshing network info cache for port 3670ef9d-bf2d-4638-a691-74ed5d418a05 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 606.404453] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-639a1f16-494a-46ef-8f77-479cbfcbbf36 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.417537] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea696dc0-45f5-498c-bbf2-4d08de9e3ac8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.446756] env[63531]: WARNING nova.virt.vmwareapi.vmops [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 01a1b3e0-4ad9-4350-b59e-499f1b3412a5 could not be found. [ 606.447032] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 606.447372] env[63531]: INFO nova.compute.manager [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] Took 0.05 seconds to destroy the instance on the hypervisor. [ 606.447609] env[63531]: DEBUG oslo.service.loopingcall [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 606.450224] env[63531]: DEBUG nova.compute.manager [-] [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 606.450476] env[63531]: DEBUG nova.network.neutron [-] [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 606.484599] env[63531]: DEBUG nova.network.neutron [-] [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 606.675149] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e3b8c13-f70e-46f9-a0e6-aae6b17ba585 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.684488] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34a6d2c6-d4ad-44f5-969b-f7489c8b57a0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.722279] env[63531]: DEBUG nova.compute.manager [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 606.725763] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6829143-1f9d-43fd-bf7a-44a7f6bc4136 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.733825] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59d23cca-a453-4cb8-b61d-470d554a9ed0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.749560] env[63531]: DEBUG nova.compute.provider_tree [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 606.760852] env[63531]: DEBUG nova.virt.hardware [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 606.761123] env[63531]: DEBUG nova.virt.hardware [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 606.761296] env[63531]: DEBUG nova.virt.hardware [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 606.761484] env[63531]: DEBUG nova.virt.hardware [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 606.761769] env[63531]: DEBUG nova.virt.hardware [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 606.761769] env[63531]: DEBUG nova.virt.hardware [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 606.761982] env[63531]: DEBUG nova.virt.hardware [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 606.762171] env[63531]: DEBUG nova.virt.hardware [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 606.762324] env[63531]: DEBUG nova.virt.hardware [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 606.762487] env[63531]: DEBUG nova.virt.hardware [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 606.762667] env[63531]: DEBUG nova.virt.hardware [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 606.763611] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14b2625e-91fc-4c24-84ee-fd9b656a5df2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.772038] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a28e9091-70d2-4fa9-9fa2-decafcacbd76 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.937243] env[63531]: DEBUG nova.network.neutron [req-400849a6-28be-4837-9eb1-19052e79782d req-42584587-91e7-4e21-88b5-d5cfef1b2142 service nova] [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 606.990112] env[63531]: DEBUG nova.network.neutron [-] [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.254542] env[63531]: DEBUG nova.scheduler.client.report [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 607.261991] env[63531]: DEBUG nova.network.neutron [req-400849a6-28be-4837-9eb1-19052e79782d req-42584587-91e7-4e21-88b5-d5cfef1b2142 service nova] [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.492872] env[63531]: INFO nova.compute.manager [-] [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] Took 1.04 seconds to deallocate network for instance. [ 607.496293] env[63531]: DEBUG nova.compute.claims [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] Aborting claim: {{(pid=63531) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 607.496293] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 607.763900] env[63531]: DEBUG oslo_concurrency.lockutils [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.072s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 607.763900] env[63531]: ERROR nova.compute.manager [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9608e6f9-9a3e-431c-a11d-02f74c26ee7e, please check neutron logs for more information. [ 607.763900] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] Traceback (most recent call last): [ 607.763900] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 607.763900] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] self.driver.spawn(context, instance, image_meta, [ 607.763900] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 607.763900] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 607.763900] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 607.763900] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] vm_ref = self.build_virtual_machine(instance, [ 607.764307] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 607.764307] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] vif_infos = vmwarevif.get_vif_info(self._session, [ 607.764307] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 607.764307] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] for vif in network_info: [ 607.764307] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 607.764307] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] return self._sync_wrapper(fn, *args, **kwargs) [ 607.764307] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 607.764307] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] self.wait() [ 607.764307] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 607.764307] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] self[:] = self._gt.wait() [ 607.764307] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 607.764307] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] return self._exit_event.wait() [ 607.764307] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 607.764676] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] result = hub.switch() [ 607.764676] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 607.764676] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] return self.greenlet.switch() [ 607.764676] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 607.764676] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] result = function(*args, **kwargs) [ 607.764676] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 607.764676] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] return func(*args, **kwargs) [ 607.764676] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 607.764676] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] raise e [ 607.764676] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 607.764676] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] nwinfo = self.network_api.allocate_for_instance( [ 607.764676] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 607.764676] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] created_port_ids = self._update_ports_for_instance( [ 607.765065] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 607.765065] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] with excutils.save_and_reraise_exception(): [ 607.765065] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 607.765065] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] self.force_reraise() [ 607.765065] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 607.765065] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] raise self.value [ 607.765065] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 607.765065] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] updated_port = self._update_port( [ 607.765065] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 607.765065] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] _ensure_no_port_binding_failure(port) [ 607.765065] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 607.765065] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] raise exception.PortBindingFailed(port_id=port['id']) [ 607.765426] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] nova.exception.PortBindingFailed: Binding failed for port 9608e6f9-9a3e-431c-a11d-02f74c26ee7e, please check neutron logs for more information. [ 607.765426] env[63531]: ERROR nova.compute.manager [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] [ 607.765608] env[63531]: DEBUG nova.compute.utils [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] Binding failed for port 9608e6f9-9a3e-431c-a11d-02f74c26ee7e, please check neutron logs for more information. {{(pid=63531) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 607.768811] env[63531]: DEBUG oslo_concurrency.lockutils [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.957s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 607.770802] env[63531]: INFO nova.compute.claims [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 607.773600] env[63531]: DEBUG oslo_concurrency.lockutils [req-400849a6-28be-4837-9eb1-19052e79782d req-42584587-91e7-4e21-88b5-d5cfef1b2142 service nova] Releasing lock "refresh_cache-01a1b3e0-4ad9-4350-b59e-499f1b3412a5" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 607.777723] env[63531]: DEBUG nova.compute.manager [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] Build of instance 2ae517ae-a159-4dc6-8100-457a4cea1aeb was re-scheduled: Binding failed for port 9608e6f9-9a3e-431c-a11d-02f74c26ee7e, please check neutron logs for more information. {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 607.778258] env[63531]: DEBUG nova.compute.manager [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] Unplugging VIFs for instance {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 607.778483] env[63531]: DEBUG oslo_concurrency.lockutils [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] Acquiring lock "refresh_cache-2ae517ae-a159-4dc6-8100-457a4cea1aeb" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 607.778623] env[63531]: DEBUG oslo_concurrency.lockutils [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] Acquired lock "refresh_cache-2ae517ae-a159-4dc6-8100-457a4cea1aeb" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 607.778777] env[63531]: DEBUG nova.network.neutron [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 607.953867] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Acquiring lock "5e9042b3-4dc2-4fa3-a664-c4b49a22e400" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 607.954104] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Lock "5e9042b3-4dc2-4fa3-a664-c4b49a22e400" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 608.106207] env[63531]: DEBUG nova.compute.manager [req-86acd41c-73d8-4044-bd4f-eca5136506d8 req-e639ce05-9119-40a3-a792-ba28a0d6bbbe service nova] [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] Received event network-vif-deleted-3670ef9d-bf2d-4638-a691-74ed5d418a05 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 608.128601] env[63531]: ERROR nova.compute.manager [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c30c9653-38f5-42cc-bda4-656ad1a333fc, please check neutron logs for more information. [ 608.128601] env[63531]: ERROR nova.compute.manager Traceback (most recent call last): [ 608.128601] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 608.128601] env[63531]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 608.128601] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 608.128601] env[63531]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 608.128601] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 608.128601] env[63531]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 608.128601] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 608.128601] env[63531]: ERROR nova.compute.manager self.force_reraise() [ 608.128601] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 608.128601] env[63531]: ERROR nova.compute.manager raise self.value [ 608.128601] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 608.128601] env[63531]: ERROR nova.compute.manager updated_port = self._update_port( [ 608.128601] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 608.128601] env[63531]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 608.129129] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 608.129129] env[63531]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 608.129129] env[63531]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c30c9653-38f5-42cc-bda4-656ad1a333fc, please check neutron logs for more information. [ 608.129129] env[63531]: ERROR nova.compute.manager [ 608.129129] env[63531]: Traceback (most recent call last): [ 608.129129] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 608.129129] env[63531]: listener.cb(fileno) [ 608.129129] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 608.129129] env[63531]: result = function(*args, **kwargs) [ 608.129129] env[63531]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 608.129129] env[63531]: return func(*args, **kwargs) [ 608.129129] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 608.129129] env[63531]: raise e [ 608.129129] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 608.129129] env[63531]: nwinfo = self.network_api.allocate_for_instance( [ 608.129129] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 608.129129] env[63531]: created_port_ids = self._update_ports_for_instance( [ 608.129129] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 608.129129] env[63531]: with excutils.save_and_reraise_exception(): [ 608.129129] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 608.129129] env[63531]: self.force_reraise() [ 608.129129] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 608.129129] env[63531]: raise self.value [ 608.129129] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 608.129129] env[63531]: updated_port = self._update_port( [ 608.129129] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 608.129129] env[63531]: _ensure_no_port_binding_failure(port) [ 608.129129] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 608.129129] env[63531]: raise exception.PortBindingFailed(port_id=port['id']) [ 608.130011] env[63531]: nova.exception.PortBindingFailed: Binding failed for port c30c9653-38f5-42cc-bda4-656ad1a333fc, please check neutron logs for more information. [ 608.130011] env[63531]: Removing descriptor: 17 [ 608.130011] env[63531]: ERROR nova.compute.manager [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c30c9653-38f5-42cc-bda4-656ad1a333fc, please check neutron logs for more information. [ 608.130011] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] Traceback (most recent call last): [ 608.130011] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 608.130011] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] yield resources [ 608.130011] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 608.130011] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] self.driver.spawn(context, instance, image_meta, [ 608.130011] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 608.130011] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] self._vmops.spawn(context, instance, image_meta, injected_files, [ 608.130011] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 608.130011] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] vm_ref = self.build_virtual_machine(instance, [ 608.130405] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 608.130405] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] vif_infos = vmwarevif.get_vif_info(self._session, [ 608.130405] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 608.130405] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] for vif in network_info: [ 608.130405] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 608.130405] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] return self._sync_wrapper(fn, *args, **kwargs) [ 608.130405] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 608.130405] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] self.wait() [ 608.130405] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 608.130405] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] self[:] = self._gt.wait() [ 608.130405] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 608.130405] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] return self._exit_event.wait() [ 608.130405] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 608.130817] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] result = hub.switch() [ 608.130817] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 608.130817] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] return self.greenlet.switch() [ 608.130817] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 608.130817] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] result = function(*args, **kwargs) [ 608.130817] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 608.130817] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] return func(*args, **kwargs) [ 608.130817] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 608.130817] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] raise e [ 608.130817] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 608.130817] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] nwinfo = self.network_api.allocate_for_instance( [ 608.130817] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 608.130817] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] created_port_ids = self._update_ports_for_instance( [ 608.131231] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 608.131231] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] with excutils.save_and_reraise_exception(): [ 608.131231] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 608.131231] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] self.force_reraise() [ 608.131231] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 608.131231] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] raise self.value [ 608.131231] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 608.131231] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] updated_port = self._update_port( [ 608.131231] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 608.131231] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] _ensure_no_port_binding_failure(port) [ 608.131231] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 608.131231] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] raise exception.PortBindingFailed(port_id=port['id']) [ 608.131634] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] nova.exception.PortBindingFailed: Binding failed for port c30c9653-38f5-42cc-bda4-656ad1a333fc, please check neutron logs for more information. [ 608.131634] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] [ 608.131634] env[63531]: INFO nova.compute.manager [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] Terminating instance [ 608.133747] env[63531]: DEBUG oslo_concurrency.lockutils [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] Acquiring lock "refresh_cache-8ac0012a-0855-4b73-a1e4-0a68a1336404" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 608.133975] env[63531]: DEBUG oslo_concurrency.lockutils [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] Acquired lock "refresh_cache-8ac0012a-0855-4b73-a1e4-0a68a1336404" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 608.134196] env[63531]: DEBUG nova.network.neutron [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 608.306728] env[63531]: DEBUG nova.network.neutron [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 608.362266] env[63531]: DEBUG nova.network.neutron [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 608.655785] env[63531]: DEBUG nova.network.neutron [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 608.771949] env[63531]: DEBUG nova.network.neutron [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 608.866238] env[63531]: DEBUG oslo_concurrency.lockutils [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] Releasing lock "refresh_cache-2ae517ae-a159-4dc6-8100-457a4cea1aeb" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 608.866547] env[63531]: DEBUG nova.compute.manager [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 608.867233] env[63531]: DEBUG nova.compute.manager [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 608.867468] env[63531]: DEBUG nova.network.neutron [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 608.888070] env[63531]: DEBUG nova.network.neutron [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 609.280058] env[63531]: DEBUG oslo_concurrency.lockutils [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] Releasing lock "refresh_cache-8ac0012a-0855-4b73-a1e4-0a68a1336404" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 609.281129] env[63531]: DEBUG nova.compute.manager [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 609.281129] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 609.281129] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7c7d63a7-3047-4e4d-8e87-1392fd4b22c8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.284985] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d46dc94d-a4ae-4d6d-afe4-82c55f4a8afb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.292925] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac704769-e74f-46fd-b2a9-071eda4cb7f7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.308699] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cce760d4-9df3-4537-9f37-88e784d351a3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.346627] env[63531]: WARNING nova.virt.vmwareapi.vmops [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8ac0012a-0855-4b73-a1e4-0a68a1336404 could not be found. [ 609.346931] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 609.347042] env[63531]: INFO nova.compute.manager [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] Took 0.07 seconds to destroy the instance on the hypervisor. [ 609.347332] env[63531]: DEBUG oslo.service.loopingcall [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 609.347756] env[63531]: DEBUG nova.compute.manager [-] [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 609.347942] env[63531]: DEBUG nova.network.neutron [-] [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 609.350223] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81e64f83-8e89-4a90-ad45-aa67972311e4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.358389] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6adb454f-68de-4c55-9126-b5cfa5d845e1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.374396] env[63531]: DEBUG nova.compute.provider_tree [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 609.380227] env[63531]: DEBUG nova.network.neutron [-] [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 609.394642] env[63531]: DEBUG nova.network.neutron [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 609.878641] env[63531]: DEBUG nova.scheduler.client.report [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 609.884103] env[63531]: DEBUG nova.network.neutron [-] [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 609.897534] env[63531]: INFO nova.compute.manager [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] [instance: 2ae517ae-a159-4dc6-8100-457a4cea1aeb] Took 1.03 seconds to deallocate network for instance. [ 610.215822] env[63531]: DEBUG nova.compute.manager [req-6c4c3f03-8acc-4a1d-b268-e5e060d53449 req-1c0bb4fa-c6a2-4d8d-be71-8e4cc5b0441e service nova] [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] Received event network-changed-c30c9653-38f5-42cc-bda4-656ad1a333fc {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 610.219105] env[63531]: DEBUG nova.compute.manager [req-6c4c3f03-8acc-4a1d-b268-e5e060d53449 req-1c0bb4fa-c6a2-4d8d-be71-8e4cc5b0441e service nova] [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] Refreshing instance network info cache due to event network-changed-c30c9653-38f5-42cc-bda4-656ad1a333fc. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 610.220416] env[63531]: DEBUG oslo_concurrency.lockutils [req-6c4c3f03-8acc-4a1d-b268-e5e060d53449 req-1c0bb4fa-c6a2-4d8d-be71-8e4cc5b0441e service nova] Acquiring lock "refresh_cache-8ac0012a-0855-4b73-a1e4-0a68a1336404" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 610.220416] env[63531]: DEBUG oslo_concurrency.lockutils [req-6c4c3f03-8acc-4a1d-b268-e5e060d53449 req-1c0bb4fa-c6a2-4d8d-be71-8e4cc5b0441e service nova] Acquired lock "refresh_cache-8ac0012a-0855-4b73-a1e4-0a68a1336404" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 610.220416] env[63531]: DEBUG nova.network.neutron [req-6c4c3f03-8acc-4a1d-b268-e5e060d53449 req-1c0bb4fa-c6a2-4d8d-be71-8e4cc5b0441e service nova] [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] Refreshing network info cache for port c30c9653-38f5-42cc-bda4-656ad1a333fc {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 610.385315] env[63531]: DEBUG oslo_concurrency.lockutils [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.615s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 610.385315] env[63531]: DEBUG nova.compute.manager [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 610.387311] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.350s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 610.389428] env[63531]: INFO nova.compute.claims [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 610.392438] env[63531]: INFO nova.compute.manager [-] [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] Took 1.04 seconds to deallocate network for instance. [ 610.394934] env[63531]: DEBUG nova.compute.claims [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] Aborting claim: {{(pid=63531) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 610.395244] env[63531]: DEBUG oslo_concurrency.lockutils [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 610.560744] env[63531]: DEBUG oslo_concurrency.lockutils [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Acquiring lock "c3920e60-b27f-496b-b616-afb23ab0bc26" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 610.561069] env[63531]: DEBUG oslo_concurrency.lockutils [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Lock "c3920e60-b27f-496b-b616-afb23ab0bc26" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 610.748366] env[63531]: DEBUG nova.network.neutron [req-6c4c3f03-8acc-4a1d-b268-e5e060d53449 req-1c0bb4fa-c6a2-4d8d-be71-8e4cc5b0441e service nova] [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 610.876392] env[63531]: DEBUG nova.network.neutron [req-6c4c3f03-8acc-4a1d-b268-e5e060d53449 req-1c0bb4fa-c6a2-4d8d-be71-8e4cc5b0441e service nova] [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.895447] env[63531]: DEBUG nova.compute.utils [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 610.900076] env[63531]: DEBUG nova.compute.manager [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 610.900264] env[63531]: DEBUG nova.network.neutron [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 610.940469] env[63531]: INFO nova.scheduler.client.report [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] Deleted allocations for instance 2ae517ae-a159-4dc6-8100-457a4cea1aeb [ 610.979832] env[63531]: DEBUG nova.policy [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '694d4b45031442b8be59b893518a0a42', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2de66f356529416d8de44e672c8f0aa0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 611.380962] env[63531]: DEBUG oslo_concurrency.lockutils [req-6c4c3f03-8acc-4a1d-b268-e5e060d53449 req-1c0bb4fa-c6a2-4d8d-be71-8e4cc5b0441e service nova] Releasing lock "refresh_cache-8ac0012a-0855-4b73-a1e4-0a68a1336404" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 611.381076] env[63531]: DEBUG nova.compute.manager [req-6c4c3f03-8acc-4a1d-b268-e5e060d53449 req-1c0bb4fa-c6a2-4d8d-be71-8e4cc5b0441e service nova] [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] Received event network-vif-deleted-c30c9653-38f5-42cc-bda4-656ad1a333fc {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 611.411890] env[63531]: DEBUG nova.compute.manager [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 611.447191] env[63531]: DEBUG oslo_concurrency.lockutils [None req-38e82583-7c8f-4759-a2b7-43619980cd3b tempest-ServerExternalEventsTest-917303244 tempest-ServerExternalEventsTest-917303244-project-member] Lock "2ae517ae-a159-4dc6-8100-457a4cea1aeb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 55.427s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 611.630120] env[63531]: DEBUG nova.network.neutron [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] Successfully created port: c90c3150-57cf-43ab-9ea6-c88af929ea4c {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 611.950148] env[63531]: DEBUG nova.compute.manager [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 611.986192] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c0b69a7-1e34-4209-9c65-6af9e4e9c678 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.995518] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3aad69d-3168-425c-8569-04820c42abb6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.026999] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8984bb25-aab8-4a61-a8a6-3310b9ab8dd8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.039339] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-296abe6e-94b4-40b7-affe-c3c9b2add018 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.049865] env[63531]: DEBUG nova.compute.provider_tree [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 612.419560] env[63531]: DEBUG nova.compute.manager [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 612.449889] env[63531]: DEBUG nova.virt.hardware [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 612.449889] env[63531]: DEBUG nova.virt.hardware [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 612.449889] env[63531]: DEBUG nova.virt.hardware [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 612.450091] env[63531]: DEBUG nova.virt.hardware [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 612.450091] env[63531]: DEBUG nova.virt.hardware [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 612.450165] env[63531]: DEBUG nova.virt.hardware [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 612.450347] env[63531]: DEBUG nova.virt.hardware [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 612.450508] env[63531]: DEBUG nova.virt.hardware [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 612.450674] env[63531]: DEBUG nova.virt.hardware [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 612.450835] env[63531]: DEBUG nova.virt.hardware [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 612.451044] env[63531]: DEBUG nova.virt.hardware [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 612.451934] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09e3b42b-b330-4003-bb11-3cd2a1aea5ec {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.462734] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d494d3fc-07ea-4b41-a2c4-636d79e572cc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.479205] env[63531]: DEBUG oslo_concurrency.lockutils [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 612.553510] env[63531]: DEBUG nova.scheduler.client.report [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 613.036272] env[63531]: DEBUG nova.compute.manager [req-6708590e-f5c1-4777-afb6-011329f915a2 req-70df3cdb-81be-4c8f-bb75-da60a346da4b service nova] [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] Received event network-changed-c90c3150-57cf-43ab-9ea6-c88af929ea4c {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 613.036558] env[63531]: DEBUG nova.compute.manager [req-6708590e-f5c1-4777-afb6-011329f915a2 req-70df3cdb-81be-4c8f-bb75-da60a346da4b service nova] [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] Refreshing instance network info cache due to event network-changed-c90c3150-57cf-43ab-9ea6-c88af929ea4c. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 613.036683] env[63531]: DEBUG oslo_concurrency.lockutils [req-6708590e-f5c1-4777-afb6-011329f915a2 req-70df3cdb-81be-4c8f-bb75-da60a346da4b service nova] Acquiring lock "refresh_cache-3a02946a-64e9-41b7-95d7-0b32766d735c" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 613.036791] env[63531]: DEBUG oslo_concurrency.lockutils [req-6708590e-f5c1-4777-afb6-011329f915a2 req-70df3cdb-81be-4c8f-bb75-da60a346da4b service nova] Acquired lock "refresh_cache-3a02946a-64e9-41b7-95d7-0b32766d735c" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 613.036942] env[63531]: DEBUG nova.network.neutron [req-6708590e-f5c1-4777-afb6-011329f915a2 req-70df3cdb-81be-4c8f-bb75-da60a346da4b service nova] [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] Refreshing network info cache for port c90c3150-57cf-43ab-9ea6-c88af929ea4c {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 613.058563] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.671s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 613.061386] env[63531]: DEBUG nova.compute.manager [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 613.067021] env[63531]: DEBUG oslo_concurrency.lockutils [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.510s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 613.067021] env[63531]: INFO nova.compute.claims [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 613.134201] env[63531]: ERROR nova.compute.manager [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c90c3150-57cf-43ab-9ea6-c88af929ea4c, please check neutron logs for more information. [ 613.134201] env[63531]: ERROR nova.compute.manager Traceback (most recent call last): [ 613.134201] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 613.134201] env[63531]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 613.134201] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 613.134201] env[63531]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 613.134201] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 613.134201] env[63531]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 613.134201] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 613.134201] env[63531]: ERROR nova.compute.manager self.force_reraise() [ 613.134201] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 613.134201] env[63531]: ERROR nova.compute.manager raise self.value [ 613.134201] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 613.134201] env[63531]: ERROR nova.compute.manager updated_port = self._update_port( [ 613.134201] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 613.134201] env[63531]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 613.134758] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 613.134758] env[63531]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 613.134758] env[63531]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c90c3150-57cf-43ab-9ea6-c88af929ea4c, please check neutron logs for more information. [ 613.134758] env[63531]: ERROR nova.compute.manager [ 613.134758] env[63531]: Traceback (most recent call last): [ 613.134758] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 613.134758] env[63531]: listener.cb(fileno) [ 613.134758] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 613.134758] env[63531]: result = function(*args, **kwargs) [ 613.134758] env[63531]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 613.134758] env[63531]: return func(*args, **kwargs) [ 613.134758] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 613.134758] env[63531]: raise e [ 613.134758] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 613.134758] env[63531]: nwinfo = self.network_api.allocate_for_instance( [ 613.134758] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 613.134758] env[63531]: created_port_ids = self._update_ports_for_instance( [ 613.134758] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 613.134758] env[63531]: with excutils.save_and_reraise_exception(): [ 613.134758] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 613.134758] env[63531]: self.force_reraise() [ 613.134758] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 613.134758] env[63531]: raise self.value [ 613.134758] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 613.134758] env[63531]: updated_port = self._update_port( [ 613.134758] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 613.134758] env[63531]: _ensure_no_port_binding_failure(port) [ 613.134758] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 613.134758] env[63531]: raise exception.PortBindingFailed(port_id=port['id']) [ 613.135683] env[63531]: nova.exception.PortBindingFailed: Binding failed for port c90c3150-57cf-43ab-9ea6-c88af929ea4c, please check neutron logs for more information. [ 613.135683] env[63531]: Removing descriptor: 17 [ 613.135683] env[63531]: ERROR nova.compute.manager [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c90c3150-57cf-43ab-9ea6-c88af929ea4c, please check neutron logs for more information. [ 613.135683] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] Traceback (most recent call last): [ 613.135683] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 613.135683] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] yield resources [ 613.135683] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 613.135683] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] self.driver.spawn(context, instance, image_meta, [ 613.135683] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 613.135683] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 613.135683] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 613.135683] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] vm_ref = self.build_virtual_machine(instance, [ 613.136072] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 613.136072] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] vif_infos = vmwarevif.get_vif_info(self._session, [ 613.136072] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 613.136072] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] for vif in network_info: [ 613.136072] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 613.136072] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] return self._sync_wrapper(fn, *args, **kwargs) [ 613.136072] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 613.136072] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] self.wait() [ 613.136072] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 613.136072] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] self[:] = self._gt.wait() [ 613.136072] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 613.136072] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] return self._exit_event.wait() [ 613.136072] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 613.136745] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] result = hub.switch() [ 613.136745] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 613.136745] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] return self.greenlet.switch() [ 613.136745] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 613.136745] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] result = function(*args, **kwargs) [ 613.136745] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 613.136745] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] return func(*args, **kwargs) [ 613.136745] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 613.136745] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] raise e [ 613.136745] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 613.136745] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] nwinfo = self.network_api.allocate_for_instance( [ 613.136745] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 613.136745] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] created_port_ids = self._update_ports_for_instance( [ 613.137196] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 613.137196] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] with excutils.save_and_reraise_exception(): [ 613.137196] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 613.137196] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] self.force_reraise() [ 613.137196] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 613.137196] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] raise self.value [ 613.137196] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 613.137196] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] updated_port = self._update_port( [ 613.137196] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 613.137196] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] _ensure_no_port_binding_failure(port) [ 613.137196] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 613.137196] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] raise exception.PortBindingFailed(port_id=port['id']) [ 613.137787] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] nova.exception.PortBindingFailed: Binding failed for port c90c3150-57cf-43ab-9ea6-c88af929ea4c, please check neutron logs for more information. [ 613.137787] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] [ 613.137787] env[63531]: INFO nova.compute.manager [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] Terminating instance [ 613.142901] env[63531]: DEBUG oslo_concurrency.lockutils [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Acquiring lock "refresh_cache-3a02946a-64e9-41b7-95d7-0b32766d735c" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 613.565275] env[63531]: DEBUG nova.network.neutron [req-6708590e-f5c1-4777-afb6-011329f915a2 req-70df3cdb-81be-4c8f-bb75-da60a346da4b service nova] [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 613.568341] env[63531]: DEBUG nova.compute.utils [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 613.572259] env[63531]: DEBUG nova.compute.manager [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 613.576259] env[63531]: DEBUG nova.network.neutron [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 613.711269] env[63531]: DEBUG nova.policy [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '95eb684ab27147f79fd7ef6ea515555f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0caca7e3c71f4ae7b9f6b8d8e7c874a8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 613.772173] env[63531]: DEBUG nova.network.neutron [req-6708590e-f5c1-4777-afb6-011329f915a2 req-70df3cdb-81be-4c8f-bb75-da60a346da4b service nova] [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 614.089547] env[63531]: DEBUG nova.compute.manager [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 614.277053] env[63531]: DEBUG oslo_concurrency.lockutils [req-6708590e-f5c1-4777-afb6-011329f915a2 req-70df3cdb-81be-4c8f-bb75-da60a346da4b service nova] Releasing lock "refresh_cache-3a02946a-64e9-41b7-95d7-0b32766d735c" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 614.278040] env[63531]: DEBUG oslo_concurrency.lockutils [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Acquired lock "refresh_cache-3a02946a-64e9-41b7-95d7-0b32766d735c" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 614.278134] env[63531]: DEBUG nova.network.neutron [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 614.384622] env[63531]: DEBUG nova.network.neutron [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] Successfully created port: 62dcd413-b725-4384-ba5a-9ccaffb40cd3 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 614.557025] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fae97d7-d086-4aaf-91f7-43b82c639af7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.565887] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84dc922a-0990-4a68-b5bb-b2bdc10b3c4b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.604194] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9b80cca-71f5-4152-b79f-0248cba5a51a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.613368] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f265d52a-1e8a-4990-95db-d6685b044a75 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.627849] env[63531]: DEBUG nova.compute.provider_tree [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 614.824677] env[63531]: DEBUG nova.network.neutron [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 615.109629] env[63531]: DEBUG nova.compute.manager [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 615.133831] env[63531]: DEBUG nova.scheduler.client.report [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 615.143536] env[63531]: DEBUG nova.virt.hardware [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 615.143721] env[63531]: DEBUG nova.virt.hardware [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 615.143867] env[63531]: DEBUG nova.virt.hardware [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 615.144068] env[63531]: DEBUG nova.virt.hardware [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 615.144439] env[63531]: DEBUG nova.virt.hardware [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 615.144733] env[63531]: DEBUG nova.virt.hardware [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 615.144835] env[63531]: DEBUG nova.virt.hardware [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 615.144996] env[63531]: DEBUG nova.virt.hardware [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 615.145709] env[63531]: DEBUG nova.virt.hardware [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 615.145709] env[63531]: DEBUG nova.virt.hardware [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 615.145709] env[63531]: DEBUG nova.virt.hardware [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 615.146850] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2f74698-8adb-4d7f-a722-74e3207b1ec0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.162567] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9846a7cc-f1e1-4833-9366-95feaf85bb95 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.220904] env[63531]: DEBUG nova.network.neutron [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 615.317395] env[63531]: DEBUG nova.compute.manager [req-09ccb1b1-864c-43bd-940f-d1066d881e2a req-9ae1f143-0cca-4a5f-9d74-fd3d4df0b9e1 service nova] [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] Received event network-vif-deleted-c90c3150-57cf-43ab-9ea6-c88af929ea4c {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 615.416899] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Acquiring lock "410ddf7a-e45b-4769-a3db-5363ce2096a7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 615.417169] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Lock "410ddf7a-e45b-4769-a3db-5363ce2096a7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 615.647715] env[63531]: DEBUG oslo_concurrency.lockutils [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.583s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 615.648249] env[63531]: DEBUG nova.compute.manager [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 615.653921] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a8d3fe75-9fbe-4167-b227-285d91eb9660 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.853s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 615.653921] env[63531]: DEBUG nova.objects.instance [None req-a8d3fe75-9fbe-4167-b227-285d91eb9660 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Lazy-loading 'resources' on Instance uuid 5b86b63a-f6c1-4983-a3cc-3cde7a10d17b {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 615.724580] env[63531]: DEBUG oslo_concurrency.lockutils [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Releasing lock "refresh_cache-3a02946a-64e9-41b7-95d7-0b32766d735c" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 615.724580] env[63531]: DEBUG nova.compute.manager [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 615.724580] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 615.724580] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4ea2f856-fbeb-4ad2-bedc-6c84271ce092 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.736336] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c20c4d3b-b7cb-40da-bb25-90e9b38f21ab {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.759797] env[63531]: WARNING nova.virt.vmwareapi.vmops [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3a02946a-64e9-41b7-95d7-0b32766d735c could not be found. [ 615.760053] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 615.760263] env[63531]: INFO nova.compute.manager [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] Took 0.04 seconds to destroy the instance on the hypervisor. [ 615.760486] env[63531]: DEBUG oslo.service.loopingcall [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 615.760721] env[63531]: DEBUG nova.compute.manager [-] [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 615.760821] env[63531]: DEBUG nova.network.neutron [-] [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 615.792547] env[63531]: DEBUG nova.network.neutron [-] [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 616.159097] env[63531]: DEBUG nova.compute.utils [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 616.162184] env[63531]: DEBUG nova.compute.manager [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 616.162363] env[63531]: DEBUG nova.network.neutron [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 616.265599] env[63531]: DEBUG nova.policy [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e6b3c98c0df5444ba7956b415b1f4e0d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ca37b8056a734639981272b98a2c8e28', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 616.296679] env[63531]: DEBUG nova.network.neutron [-] [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 616.649696] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0c4d16b-4803-4a4a-a4a3-60ad6064d9cb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.658602] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28b3aa80-bdd7-4de2-b165-674cc9510fb1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.663216] env[63531]: DEBUG nova.compute.manager [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 616.698967] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-392efc92-c37d-4dc6-a345-5844b11fffa4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.708165] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ff3c3cb-bcdd-412b-aa44-c3a97f74fdb0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.727078] env[63531]: DEBUG nova.compute.provider_tree [None req-a8d3fe75-9fbe-4167-b227-285d91eb9660 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 616.798997] env[63531]: INFO nova.compute.manager [-] [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] Took 1.04 seconds to deallocate network for instance. [ 616.800535] env[63531]: DEBUG nova.compute.claims [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] Aborting claim: {{(pid=63531) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 616.800716] env[63531]: DEBUG oslo_concurrency.lockutils [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 616.921759] env[63531]: DEBUG nova.network.neutron [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] Successfully created port: 584241fa-8b49-462d-8c0f-0accc2a17497 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 617.016600] env[63531]: ERROR nova.compute.manager [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 62dcd413-b725-4384-ba5a-9ccaffb40cd3, please check neutron logs for more information. [ 617.016600] env[63531]: ERROR nova.compute.manager Traceback (most recent call last): [ 617.016600] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 617.016600] env[63531]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 617.016600] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 617.016600] env[63531]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 617.016600] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 617.016600] env[63531]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 617.016600] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 617.016600] env[63531]: ERROR nova.compute.manager self.force_reraise() [ 617.016600] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 617.016600] env[63531]: ERROR nova.compute.manager raise self.value [ 617.016600] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 617.016600] env[63531]: ERROR nova.compute.manager updated_port = self._update_port( [ 617.016600] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 617.016600] env[63531]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 617.017230] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 617.017230] env[63531]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 617.017230] env[63531]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 62dcd413-b725-4384-ba5a-9ccaffb40cd3, please check neutron logs for more information. [ 617.017230] env[63531]: ERROR nova.compute.manager [ 617.017230] env[63531]: Traceback (most recent call last): [ 617.017230] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 617.017230] env[63531]: listener.cb(fileno) [ 617.017230] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 617.017230] env[63531]: result = function(*args, **kwargs) [ 617.017230] env[63531]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 617.017230] env[63531]: return func(*args, **kwargs) [ 617.017230] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 617.017230] env[63531]: raise e [ 617.017230] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 617.017230] env[63531]: nwinfo = self.network_api.allocate_for_instance( [ 617.017230] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 617.017230] env[63531]: created_port_ids = self._update_ports_for_instance( [ 617.017230] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 617.017230] env[63531]: with excutils.save_and_reraise_exception(): [ 617.017230] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 617.017230] env[63531]: self.force_reraise() [ 617.017230] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 617.017230] env[63531]: raise self.value [ 617.017230] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 617.017230] env[63531]: updated_port = self._update_port( [ 617.017230] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 617.017230] env[63531]: _ensure_no_port_binding_failure(port) [ 617.017230] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 617.017230] env[63531]: raise exception.PortBindingFailed(port_id=port['id']) [ 617.018224] env[63531]: nova.exception.PortBindingFailed: Binding failed for port 62dcd413-b725-4384-ba5a-9ccaffb40cd3, please check neutron logs for more information. [ 617.018224] env[63531]: Removing descriptor: 20 [ 617.018224] env[63531]: ERROR nova.compute.manager [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 62dcd413-b725-4384-ba5a-9ccaffb40cd3, please check neutron logs for more information. [ 617.018224] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] Traceback (most recent call last): [ 617.018224] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 617.018224] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] yield resources [ 617.018224] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 617.018224] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] self.driver.spawn(context, instance, image_meta, [ 617.018224] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 617.018224] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 617.018224] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 617.018224] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] vm_ref = self.build_virtual_machine(instance, [ 617.018754] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 617.018754] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] vif_infos = vmwarevif.get_vif_info(self._session, [ 617.018754] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 617.018754] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] for vif in network_info: [ 617.018754] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 617.018754] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] return self._sync_wrapper(fn, *args, **kwargs) [ 617.018754] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 617.018754] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] self.wait() [ 617.018754] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 617.018754] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] self[:] = self._gt.wait() [ 617.018754] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 617.018754] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] return self._exit_event.wait() [ 617.018754] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 617.019247] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] result = hub.switch() [ 617.019247] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 617.019247] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] return self.greenlet.switch() [ 617.019247] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 617.019247] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] result = function(*args, **kwargs) [ 617.019247] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 617.019247] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] return func(*args, **kwargs) [ 617.019247] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 617.019247] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] raise e [ 617.019247] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 617.019247] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] nwinfo = self.network_api.allocate_for_instance( [ 617.019247] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 617.019247] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] created_port_ids = self._update_ports_for_instance( [ 617.019847] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 617.019847] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] with excutils.save_and_reraise_exception(): [ 617.019847] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 617.019847] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] self.force_reraise() [ 617.019847] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 617.019847] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] raise self.value [ 617.019847] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 617.019847] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] updated_port = self._update_port( [ 617.019847] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 617.019847] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] _ensure_no_port_binding_failure(port) [ 617.019847] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 617.019847] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] raise exception.PortBindingFailed(port_id=port['id']) [ 617.021621] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] nova.exception.PortBindingFailed: Binding failed for port 62dcd413-b725-4384-ba5a-9ccaffb40cd3, please check neutron logs for more information. [ 617.021621] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] [ 617.021621] env[63531]: INFO nova.compute.manager [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] Terminating instance [ 617.021621] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] Acquiring lock "refresh_cache-fe881dbe-6d44-4e31-a537-4aa9dfe44ff4" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 617.021621] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] Acquired lock "refresh_cache-fe881dbe-6d44-4e31-a537-4aa9dfe44ff4" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 617.021621] env[63531]: DEBUG nova.network.neutron [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 617.231880] env[63531]: DEBUG nova.scheduler.client.report [None req-a8d3fe75-9fbe-4167-b227-285d91eb9660 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 617.373550] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Acquiring lock "1ca7b851-2a06-4181-8271-58aafcd322d6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.374173] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Lock "1ca7b851-2a06-4181-8271-58aafcd322d6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 617.543053] env[63531]: DEBUG nova.network.neutron [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 617.618426] env[63531]: DEBUG nova.network.neutron [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 617.675932] env[63531]: DEBUG nova.compute.manager [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 617.712125] env[63531]: DEBUG nova.virt.hardware [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 617.712125] env[63531]: DEBUG nova.virt.hardware [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 617.712125] env[63531]: DEBUG nova.virt.hardware [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 617.712375] env[63531]: DEBUG nova.virt.hardware [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 617.712649] env[63531]: DEBUG nova.virt.hardware [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 617.712959] env[63531]: DEBUG nova.virt.hardware [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 617.713333] env[63531]: DEBUG nova.virt.hardware [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 617.713616] env[63531]: DEBUG nova.virt.hardware [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 617.713916] env[63531]: DEBUG nova.virt.hardware [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 617.714606] env[63531]: DEBUG nova.virt.hardware [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 617.715053] env[63531]: DEBUG nova.virt.hardware [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 617.717016] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-831a4785-09af-4a49-9725-5eebc354b49b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.725707] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ed4473d-82d8-4a5b-b8c3-6d09052675cf {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.748275] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a8d3fe75-9fbe-4167-b227-285d91eb9660 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.096s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 617.752907] env[63531]: DEBUG oslo_concurrency.lockutils [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 27.143s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 617.752907] env[63531]: DEBUG nova.objects.instance [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63531) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 617.790347] env[63531]: INFO nova.scheduler.client.report [None req-a8d3fe75-9fbe-4167-b227-285d91eb9660 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Deleted allocations for instance 5b86b63a-f6c1-4983-a3cc-3cde7a10d17b [ 617.982124] env[63531]: DEBUG nova.compute.manager [req-500ebdf3-cf88-4c60-92d2-6df1b4ebe7d6 req-4f047d39-a436-47b4-866b-aadaafce8b7a service nova] [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] Received event network-changed-62dcd413-b725-4384-ba5a-9ccaffb40cd3 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 617.982344] env[63531]: DEBUG nova.compute.manager [req-500ebdf3-cf88-4c60-92d2-6df1b4ebe7d6 req-4f047d39-a436-47b4-866b-aadaafce8b7a service nova] [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] Refreshing instance network info cache due to event network-changed-62dcd413-b725-4384-ba5a-9ccaffb40cd3. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 617.982536] env[63531]: DEBUG oslo_concurrency.lockutils [req-500ebdf3-cf88-4c60-92d2-6df1b4ebe7d6 req-4f047d39-a436-47b4-866b-aadaafce8b7a service nova] Acquiring lock "refresh_cache-fe881dbe-6d44-4e31-a537-4aa9dfe44ff4" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 617.998122] env[63531]: ERROR nova.compute.manager [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 584241fa-8b49-462d-8c0f-0accc2a17497, please check neutron logs for more information. [ 617.998122] env[63531]: ERROR nova.compute.manager Traceback (most recent call last): [ 617.998122] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 617.998122] env[63531]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 617.998122] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 617.998122] env[63531]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 617.998122] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 617.998122] env[63531]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 617.998122] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 617.998122] env[63531]: ERROR nova.compute.manager self.force_reraise() [ 617.998122] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 617.998122] env[63531]: ERROR nova.compute.manager raise self.value [ 617.998122] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 617.998122] env[63531]: ERROR nova.compute.manager updated_port = self._update_port( [ 617.998122] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 617.998122] env[63531]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 617.998694] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 617.998694] env[63531]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 617.998694] env[63531]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 584241fa-8b49-462d-8c0f-0accc2a17497, please check neutron logs for more information. [ 617.998694] env[63531]: ERROR nova.compute.manager [ 617.998694] env[63531]: Traceback (most recent call last): [ 617.998694] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 617.998694] env[63531]: listener.cb(fileno) [ 617.998694] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 617.998694] env[63531]: result = function(*args, **kwargs) [ 617.998694] env[63531]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 617.998694] env[63531]: return func(*args, **kwargs) [ 617.998694] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 617.998694] env[63531]: raise e [ 617.998694] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 617.998694] env[63531]: nwinfo = self.network_api.allocate_for_instance( [ 617.998694] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 617.998694] env[63531]: created_port_ids = self._update_ports_for_instance( [ 617.998694] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 617.998694] env[63531]: with excutils.save_and_reraise_exception(): [ 617.998694] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 617.998694] env[63531]: self.force_reraise() [ 617.998694] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 617.998694] env[63531]: raise self.value [ 617.998694] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 617.998694] env[63531]: updated_port = self._update_port( [ 617.998694] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 617.998694] env[63531]: _ensure_no_port_binding_failure(port) [ 617.998694] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 617.998694] env[63531]: raise exception.PortBindingFailed(port_id=port['id']) [ 617.999677] env[63531]: nova.exception.PortBindingFailed: Binding failed for port 584241fa-8b49-462d-8c0f-0accc2a17497, please check neutron logs for more information. [ 617.999677] env[63531]: Removing descriptor: 17 [ 617.999677] env[63531]: ERROR nova.compute.manager [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 584241fa-8b49-462d-8c0f-0accc2a17497, please check neutron logs for more information. [ 617.999677] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] Traceback (most recent call last): [ 617.999677] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 617.999677] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] yield resources [ 617.999677] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 617.999677] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] self.driver.spawn(context, instance, image_meta, [ 617.999677] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 617.999677] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] self._vmops.spawn(context, instance, image_meta, injected_files, [ 617.999677] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 617.999677] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] vm_ref = self.build_virtual_machine(instance, [ 618.000119] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 618.000119] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] vif_infos = vmwarevif.get_vif_info(self._session, [ 618.000119] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 618.000119] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] for vif in network_info: [ 618.000119] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 618.000119] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] return self._sync_wrapper(fn, *args, **kwargs) [ 618.000119] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 618.000119] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] self.wait() [ 618.000119] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 618.000119] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] self[:] = self._gt.wait() [ 618.000119] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 618.000119] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] return self._exit_event.wait() [ 618.000119] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 618.000499] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] result = hub.switch() [ 618.000499] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 618.000499] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] return self.greenlet.switch() [ 618.000499] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 618.000499] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] result = function(*args, **kwargs) [ 618.000499] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 618.000499] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] return func(*args, **kwargs) [ 618.000499] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 618.000499] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] raise e [ 618.000499] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 618.000499] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] nwinfo = self.network_api.allocate_for_instance( [ 618.000499] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 618.000499] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] created_port_ids = self._update_ports_for_instance( [ 618.000893] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 618.000893] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] with excutils.save_and_reraise_exception(): [ 618.000893] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 618.000893] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] self.force_reraise() [ 618.000893] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 618.000893] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] raise self.value [ 618.000893] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 618.000893] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] updated_port = self._update_port( [ 618.000893] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 618.000893] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] _ensure_no_port_binding_failure(port) [ 618.000893] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 618.000893] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] raise exception.PortBindingFailed(port_id=port['id']) [ 618.001369] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] nova.exception.PortBindingFailed: Binding failed for port 584241fa-8b49-462d-8c0f-0accc2a17497, please check neutron logs for more information. [ 618.001369] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] [ 618.001369] env[63531]: INFO nova.compute.manager [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] Terminating instance [ 618.003240] env[63531]: DEBUG oslo_concurrency.lockutils [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] Acquiring lock "refresh_cache-3c513284-4860-46b2-8ba1-e185efe0db86" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 618.003240] env[63531]: DEBUG oslo_concurrency.lockutils [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] Acquired lock "refresh_cache-3c513284-4860-46b2-8ba1-e185efe0db86" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 618.003240] env[63531]: DEBUG nova.network.neutron [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 618.121202] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] Releasing lock "refresh_cache-fe881dbe-6d44-4e31-a537-4aa9dfe44ff4" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 618.121202] env[63531]: DEBUG nova.compute.manager [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 618.121202] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 618.121577] env[63531]: DEBUG oslo_concurrency.lockutils [req-500ebdf3-cf88-4c60-92d2-6df1b4ebe7d6 req-4f047d39-a436-47b4-866b-aadaafce8b7a service nova] Acquired lock "refresh_cache-fe881dbe-6d44-4e31-a537-4aa9dfe44ff4" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 618.121788] env[63531]: DEBUG nova.network.neutron [req-500ebdf3-cf88-4c60-92d2-6df1b4ebe7d6 req-4f047d39-a436-47b4-866b-aadaafce8b7a service nova] [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] Refreshing network info cache for port 62dcd413-b725-4384-ba5a-9ccaffb40cd3 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 618.123648] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-db9c92bd-6d5c-47a5-948b-8b91748b9a5d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.136022] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da27ab60-dbae-4de2-ae40-d23e53812a3d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.161962] env[63531]: WARNING nova.virt.vmwareapi.vmops [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance fe881dbe-6d44-4e31-a537-4aa9dfe44ff4 could not be found. [ 618.162217] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 618.162402] env[63531]: INFO nova.compute.manager [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] Took 0.04 seconds to destroy the instance on the hypervisor. [ 618.162654] env[63531]: DEBUG oslo.service.loopingcall [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 618.163235] env[63531]: DEBUG nova.compute.manager [-] [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 618.163488] env[63531]: DEBUG nova.network.neutron [-] [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 618.187032] env[63531]: DEBUG nova.network.neutron [-] [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 618.304416] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a8d3fe75-9fbe-4167-b227-285d91eb9660 tempest-ServerDiagnosticsV248Test-786930347 tempest-ServerDiagnosticsV248Test-786930347-project-member] Lock "5b86b63a-f6c1-4983-a3cc-3cde7a10d17b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.523s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 618.537132] env[63531]: DEBUG nova.network.neutron [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 618.691980] env[63531]: DEBUG nova.network.neutron [-] [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 618.734603] env[63531]: DEBUG nova.network.neutron [req-500ebdf3-cf88-4c60-92d2-6df1b4ebe7d6 req-4f047d39-a436-47b4-866b-aadaafce8b7a service nova] [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 618.774494] env[63531]: DEBUG oslo_concurrency.lockutils [None req-011d3a05-a310-414a-bc69-c0fd2e792e81 tempest-ServersAdmin275Test-72815535 tempest-ServersAdmin275Test-72815535-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.021s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 618.776448] env[63531]: DEBUG oslo_concurrency.lockutils [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.842s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 618.926818] env[63531]: DEBUG nova.network.neutron [req-500ebdf3-cf88-4c60-92d2-6df1b4ebe7d6 req-4f047d39-a436-47b4-866b-aadaafce8b7a service nova] [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 618.944592] env[63531]: DEBUG nova.network.neutron [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 619.196111] env[63531]: INFO nova.compute.manager [-] [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] Took 1.03 seconds to deallocate network for instance. [ 619.196902] env[63531]: DEBUG nova.compute.claims [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] Aborting claim: {{(pid=63531) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 619.197120] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 619.435208] env[63531]: DEBUG oslo_concurrency.lockutils [req-500ebdf3-cf88-4c60-92d2-6df1b4ebe7d6 req-4f047d39-a436-47b4-866b-aadaafce8b7a service nova] Releasing lock "refresh_cache-fe881dbe-6d44-4e31-a537-4aa9dfe44ff4" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 619.435208] env[63531]: DEBUG nova.compute.manager [req-500ebdf3-cf88-4c60-92d2-6df1b4ebe7d6 req-4f047d39-a436-47b4-866b-aadaafce8b7a service nova] [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] Received event network-vif-deleted-62dcd413-b725-4384-ba5a-9ccaffb40cd3 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 619.449334] env[63531]: DEBUG oslo_concurrency.lockutils [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] Releasing lock "refresh_cache-3c513284-4860-46b2-8ba1-e185efe0db86" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 619.449538] env[63531]: DEBUG nova.compute.manager [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 619.449757] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 619.450361] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-651223da-6bcb-4405-a729-8bc724555477 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.467414] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-765b268e-5591-426c-9ba2-a21734dbf99e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.496313] env[63531]: WARNING nova.virt.vmwareapi.vmops [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3c513284-4860-46b2-8ba1-e185efe0db86 could not be found. [ 619.496719] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 619.496947] env[63531]: INFO nova.compute.manager [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] Took 0.05 seconds to destroy the instance on the hypervisor. [ 619.497223] env[63531]: DEBUG oslo.service.loopingcall [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 619.499992] env[63531]: DEBUG nova.compute.manager [-] [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 619.500042] env[63531]: DEBUG nova.network.neutron [-] [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 619.539180] env[63531]: DEBUG nova.network.neutron [-] [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 619.874595] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-172e1548-f1ec-4d1d-a68b-78cebf6c1225 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.883919] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6928fd5-5c8a-4b55-8fd8-b59c6ad1a8b0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.924782] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c34726c7-afe3-4db2-a882-e11721e9663e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.932778] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23b8d13b-6927-4eac-be57-22cd9d783b03 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.948745] env[63531]: DEBUG nova.compute.provider_tree [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 620.041970] env[63531]: DEBUG nova.network.neutron [-] [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.128533] env[63531]: DEBUG nova.compute.manager [req-9bf10fe7-6ea2-4b85-8319-b0ec8c7fd4d8 req-fa0a26cb-85b7-46a0-a67f-793c988052c9 service nova] [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] Received event network-changed-584241fa-8b49-462d-8c0f-0accc2a17497 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 620.129320] env[63531]: DEBUG nova.compute.manager [req-9bf10fe7-6ea2-4b85-8319-b0ec8c7fd4d8 req-fa0a26cb-85b7-46a0-a67f-793c988052c9 service nova] [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] Refreshing instance network info cache due to event network-changed-584241fa-8b49-462d-8c0f-0accc2a17497. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 620.129567] env[63531]: DEBUG oslo_concurrency.lockutils [req-9bf10fe7-6ea2-4b85-8319-b0ec8c7fd4d8 req-fa0a26cb-85b7-46a0-a67f-793c988052c9 service nova] Acquiring lock "refresh_cache-3c513284-4860-46b2-8ba1-e185efe0db86" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 620.130134] env[63531]: DEBUG oslo_concurrency.lockutils [req-9bf10fe7-6ea2-4b85-8319-b0ec8c7fd4d8 req-fa0a26cb-85b7-46a0-a67f-793c988052c9 service nova] Acquired lock "refresh_cache-3c513284-4860-46b2-8ba1-e185efe0db86" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 620.130134] env[63531]: DEBUG nova.network.neutron [req-9bf10fe7-6ea2-4b85-8319-b0ec8c7fd4d8 req-fa0a26cb-85b7-46a0-a67f-793c988052c9 service nova] [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] Refreshing network info cache for port 584241fa-8b49-462d-8c0f-0accc2a17497 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 620.451406] env[63531]: DEBUG nova.scheduler.client.report [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 620.549787] env[63531]: INFO nova.compute.manager [-] [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] Took 1.05 seconds to deallocate network for instance. [ 620.552052] env[63531]: DEBUG nova.compute.claims [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] Aborting claim: {{(pid=63531) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 620.552239] env[63531]: DEBUG oslo_concurrency.lockutils [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 620.673182] env[63531]: DEBUG nova.network.neutron [req-9bf10fe7-6ea2-4b85-8319-b0ec8c7fd4d8 req-fa0a26cb-85b7-46a0-a67f-793c988052c9 service nova] [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 620.785546] env[63531]: DEBUG nova.network.neutron [req-9bf10fe7-6ea2-4b85-8319-b0ec8c7fd4d8 req-fa0a26cb-85b7-46a0-a67f-793c988052c9 service nova] [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.962102] env[63531]: DEBUG oslo_concurrency.lockutils [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.185s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 620.962102] env[63531]: ERROR nova.compute.manager [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f474eb4d-6d17-448a-8665-0b00d2223b52, please check neutron logs for more information. [ 620.962102] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] Traceback (most recent call last): [ 620.962102] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 620.962102] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] self.driver.spawn(context, instance, image_meta, [ 620.962102] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 620.962102] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 620.962102] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 620.962102] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] vm_ref = self.build_virtual_machine(instance, [ 620.962408] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 620.962408] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] vif_infos = vmwarevif.get_vif_info(self._session, [ 620.962408] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 620.962408] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] for vif in network_info: [ 620.962408] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 620.962408] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] return self._sync_wrapper(fn, *args, **kwargs) [ 620.962408] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 620.962408] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] self.wait() [ 620.962408] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 620.962408] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] self[:] = self._gt.wait() [ 620.962408] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 620.962408] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] return self._exit_event.wait() [ 620.962408] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 620.962658] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] result = hub.switch() [ 620.962658] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 620.962658] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] return self.greenlet.switch() [ 620.962658] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 620.962658] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] result = function(*args, **kwargs) [ 620.962658] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 620.962658] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] return func(*args, **kwargs) [ 620.962658] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 620.962658] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] raise e [ 620.962658] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 620.962658] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] nwinfo = self.network_api.allocate_for_instance( [ 620.962658] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 620.962658] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] created_port_ids = self._update_ports_for_instance( [ 620.962912] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 620.962912] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] with excutils.save_and_reraise_exception(): [ 620.962912] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 620.962912] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] self.force_reraise() [ 620.962912] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 620.962912] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] raise self.value [ 620.962912] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 620.962912] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] updated_port = self._update_port( [ 620.962912] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 620.962912] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] _ensure_no_port_binding_failure(port) [ 620.962912] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 620.962912] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] raise exception.PortBindingFailed(port_id=port['id']) [ 620.963166] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] nova.exception.PortBindingFailed: Binding failed for port f474eb4d-6d17-448a-8665-0b00d2223b52, please check neutron logs for more information. [ 620.963166] env[63531]: ERROR nova.compute.manager [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] [ 620.963166] env[63531]: DEBUG nova.compute.utils [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] Binding failed for port f474eb4d-6d17-448a-8665-0b00d2223b52, please check neutron logs for more information. {{(pid=63531) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 620.967214] env[63531]: DEBUG nova.compute.manager [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] Build of instance 9fa373b3-8e15-4d26-82be-7dda51bc6cf6 was re-scheduled: Binding failed for port f474eb4d-6d17-448a-8665-0b00d2223b52, please check neutron logs for more information. {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 620.967214] env[63531]: DEBUG nova.compute.manager [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] Unplugging VIFs for instance {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 620.967214] env[63531]: DEBUG oslo_concurrency.lockutils [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Acquiring lock "refresh_cache-9fa373b3-8e15-4d26-82be-7dda51bc6cf6" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 620.967214] env[63531]: DEBUG oslo_concurrency.lockutils [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Acquired lock "refresh_cache-9fa373b3-8e15-4d26-82be-7dda51bc6cf6" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 620.967366] env[63531]: DEBUG nova.network.neutron [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 620.967875] env[63531]: DEBUG oslo_concurrency.lockutils [None req-da9f7a66-33b6-42ee-90cf-f8ce53a3c4c8 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.539s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 620.969055] env[63531]: DEBUG nova.objects.instance [None req-da9f7a66-33b6-42ee-90cf-f8ce53a3c4c8 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Lazy-loading 'resources' on Instance uuid 4410b013-868a-41e7-9dfb-584fb2d9d142 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 621.289541] env[63531]: DEBUG oslo_concurrency.lockutils [req-9bf10fe7-6ea2-4b85-8319-b0ec8c7fd4d8 req-fa0a26cb-85b7-46a0-a67f-793c988052c9 service nova] Releasing lock "refresh_cache-3c513284-4860-46b2-8ba1-e185efe0db86" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 621.289821] env[63531]: DEBUG nova.compute.manager [req-9bf10fe7-6ea2-4b85-8319-b0ec8c7fd4d8 req-fa0a26cb-85b7-46a0-a67f-793c988052c9 service nova] [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] Received event network-vif-deleted-584241fa-8b49-462d-8c0f-0accc2a17497 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 621.496376] env[63531]: DEBUG nova.network.neutron [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 621.619341] env[63531]: DEBUG nova.network.neutron [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.643664] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Acquiring lock "e7fb00a3-0d2e-4f54-950d-337307112d7a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 621.643908] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Lock "e7fb00a3-0d2e-4f54-950d-337307112d7a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 621.702562] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Acquiring lock "0af99240-fa7a-4eed-a729-a5ae98c41cf9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 621.703031] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Lock "0af99240-fa7a-4eed-a729-a5ae98c41cf9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 621.932064] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81bc4c38-5da4-4616-956f-257c9f6e6be9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.939454] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e259e273-e6df-4269-9bd3-0875df743b53 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.968634] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60252a40-352c-4971-990b-ffc452fd7f21 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.975278] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e60592f-74c6-4992-80b8-07c3816ee5de {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.988257] env[63531]: DEBUG nova.compute.provider_tree [None req-da9f7a66-33b6-42ee-90cf-f8ce53a3c4c8 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 622.122208] env[63531]: DEBUG oslo_concurrency.lockutils [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Releasing lock "refresh_cache-9fa373b3-8e15-4d26-82be-7dda51bc6cf6" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 622.122462] env[63531]: DEBUG nova.compute.manager [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 622.122644] env[63531]: DEBUG nova.compute.manager [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 622.122823] env[63531]: DEBUG nova.network.neutron [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 622.138171] env[63531]: DEBUG nova.network.neutron [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 622.491464] env[63531]: DEBUG nova.scheduler.client.report [None req-da9f7a66-33b6-42ee-90cf-f8ce53a3c4c8 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 622.641069] env[63531]: DEBUG nova.network.neutron [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 622.997100] env[63531]: DEBUG oslo_concurrency.lockutils [None req-da9f7a66-33b6-42ee-90cf-f8ce53a3c4c8 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.029s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 622.999613] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 22.596s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 623.016109] env[63531]: INFO nova.scheduler.client.report [None req-da9f7a66-33b6-42ee-90cf-f8ce53a3c4c8 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Deleted allocations for instance 4410b013-868a-41e7-9dfb-584fb2d9d142 [ 623.144557] env[63531]: INFO nova.compute.manager [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: 9fa373b3-8e15-4d26-82be-7dda51bc6cf6] Took 1.02 seconds to deallocate network for instance. [ 623.523538] env[63531]: DEBUG oslo_concurrency.lockutils [None req-da9f7a66-33b6-42ee-90cf-f8ce53a3c4c8 tempest-ServersAdmin275Test-1249152978 tempest-ServersAdmin275Test-1249152978-project-member] Lock "4410b013-868a-41e7-9dfb-584fb2d9d142" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.972s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 623.891867] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-999658fd-57d4-497c-a9b9-6c9064c6fa07 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.899784] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3935de80-58e4-4dfb-a330-6fd2922a3ed1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.930046] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dd28183-3b09-499f-b9a2-d579f0f51722 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.938156] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25c4eb06-ac9a-4203-a074-7f6302e8a8e3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.952771] env[63531]: DEBUG nova.compute.provider_tree [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 624.176078] env[63531]: INFO nova.scheduler.client.report [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Deleted allocations for instance 9fa373b3-8e15-4d26-82be-7dda51bc6cf6 [ 624.457092] env[63531]: DEBUG nova.scheduler.client.report [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 624.685543] env[63531]: DEBUG oslo_concurrency.lockutils [None req-33c967b8-d20b-4927-977d-7595523737ac tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Lock "9fa373b3-8e15-4d26-82be-7dda51bc6cf6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 66.735s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 624.963236] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.963s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 624.965064] env[63531]: ERROR nova.compute.manager [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] [instance: d941a593-7047-436a-b95d-bcb4850829f5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 38892f48-025c-4a63-b5ca-422b27c6177d, please check neutron logs for more information. [ 624.965064] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] Traceback (most recent call last): [ 624.965064] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 624.965064] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] self.driver.spawn(context, instance, image_meta, [ 624.965064] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 624.965064] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 624.965064] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 624.965064] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] vm_ref = self.build_virtual_machine(instance, [ 624.965064] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 624.965064] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] vif_infos = vmwarevif.get_vif_info(self._session, [ 624.965064] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 624.965372] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] for vif in network_info: [ 624.965372] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 624.965372] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] return self._sync_wrapper(fn, *args, **kwargs) [ 624.965372] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 624.965372] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] self.wait() [ 624.965372] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 624.965372] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] self[:] = self._gt.wait() [ 624.965372] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 624.965372] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] return self._exit_event.wait() [ 624.965372] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 624.965372] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] result = hub.switch() [ 624.965372] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 624.965372] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] return self.greenlet.switch() [ 624.965642] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 624.965642] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] result = function(*args, **kwargs) [ 624.965642] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 624.965642] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] return func(*args, **kwargs) [ 624.965642] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 624.965642] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] raise e [ 624.965642] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 624.965642] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] nwinfo = self.network_api.allocate_for_instance( [ 624.965642] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 624.965642] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] created_port_ids = self._update_ports_for_instance( [ 624.965642] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 624.965642] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] with excutils.save_and_reraise_exception(): [ 624.965642] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 624.965905] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] self.force_reraise() [ 624.965905] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 624.965905] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] raise self.value [ 624.965905] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 624.965905] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] updated_port = self._update_port( [ 624.965905] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 624.965905] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] _ensure_no_port_binding_failure(port) [ 624.965905] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 624.965905] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] raise exception.PortBindingFailed(port_id=port['id']) [ 624.965905] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] nova.exception.PortBindingFailed: Binding failed for port 38892f48-025c-4a63-b5ca-422b27c6177d, please check neutron logs for more information. [ 624.965905] env[63531]: ERROR nova.compute.manager [instance: d941a593-7047-436a-b95d-bcb4850829f5] [ 624.966145] env[63531]: DEBUG nova.compute.utils [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] [instance: d941a593-7047-436a-b95d-bcb4850829f5] Binding failed for port 38892f48-025c-4a63-b5ca-422b27c6177d, please check neutron logs for more information. {{(pid=63531) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 624.966775] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.085s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.968735] env[63531]: INFO nova.compute.claims [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 56872ede-96f0-48ca-b289-86784f37527d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 624.971490] env[63531]: DEBUG nova.compute.manager [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] [instance: d941a593-7047-436a-b95d-bcb4850829f5] Build of instance d941a593-7047-436a-b95d-bcb4850829f5 was re-scheduled: Binding failed for port 38892f48-025c-4a63-b5ca-422b27c6177d, please check neutron logs for more information. {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 624.971931] env[63531]: DEBUG nova.compute.manager [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] [instance: d941a593-7047-436a-b95d-bcb4850829f5] Unplugging VIFs for instance {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 624.972169] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] Acquiring lock "refresh_cache-d941a593-7047-436a-b95d-bcb4850829f5" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 624.972318] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] Acquired lock "refresh_cache-d941a593-7047-436a-b95d-bcb4850829f5" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.972477] env[63531]: DEBUG nova.network.neutron [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] [instance: d941a593-7047-436a-b95d-bcb4850829f5] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 625.188433] env[63531]: DEBUG nova.compute.manager [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 625.494300] env[63531]: DEBUG nova.network.neutron [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] [instance: d941a593-7047-436a-b95d-bcb4850829f5] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 625.588552] env[63531]: DEBUG nova.network.neutron [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] [instance: d941a593-7047-436a-b95d-bcb4850829f5] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.711817] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 625.919605] env[63531]: DEBUG oslo_concurrency.lockutils [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Acquiring lock "e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 625.920435] env[63531]: DEBUG oslo_concurrency.lockutils [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Lock "e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 626.092841] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] Releasing lock "refresh_cache-d941a593-7047-436a-b95d-bcb4850829f5" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 626.092841] env[63531]: DEBUG nova.compute.manager [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 626.092841] env[63531]: DEBUG nova.compute.manager [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] [instance: d941a593-7047-436a-b95d-bcb4850829f5] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 626.092841] env[63531]: DEBUG nova.network.neutron [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] [instance: d941a593-7047-436a-b95d-bcb4850829f5] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 626.112322] env[63531]: DEBUG nova.network.neutron [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] [instance: d941a593-7047-436a-b95d-bcb4850829f5] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 626.429226] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b80945a-8055-4215-ac2f-e51fe1b0b9d3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.437136] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91553536-a794-4c1d-b256-66f153e4d6d8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.470155] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38ac96f0-aa60-4190-ac2f-2b95f8c306b6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.480097] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49e87dde-cc96-4467-bf8f-2e26cb808389 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.494622] env[63531]: DEBUG nova.compute.provider_tree [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 626.618771] env[63531]: DEBUG nova.network.neutron [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] [instance: d941a593-7047-436a-b95d-bcb4850829f5] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.002086] env[63531]: DEBUG nova.scheduler.client.report [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 627.121648] env[63531]: INFO nova.compute.manager [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] [instance: d941a593-7047-436a-b95d-bcb4850829f5] Took 1.03 seconds to deallocate network for instance. [ 627.512253] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.543s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 627.512253] env[63531]: DEBUG nova.compute.manager [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 56872ede-96f0-48ca-b289-86784f37527d] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 627.516678] env[63531]: DEBUG oslo_concurrency.lockutils [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 22.219s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 628.026404] env[63531]: DEBUG nova.compute.utils [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 628.027903] env[63531]: DEBUG nova.compute.manager [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 56872ede-96f0-48ca-b289-86784f37527d] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 628.028665] env[63531]: DEBUG nova.network.neutron [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 56872ede-96f0-48ca-b289-86784f37527d] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 628.077413] env[63531]: DEBUG nova.policy [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '694d4b45031442b8be59b893518a0a42', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2de66f356529416d8de44e672c8f0aa0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 628.155251] env[63531]: INFO nova.scheduler.client.report [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] Deleted allocations for instance d941a593-7047-436a-b95d-bcb4850829f5 [ 628.333970] env[63531]: DEBUG nova.network.neutron [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 56872ede-96f0-48ca-b289-86784f37527d] Successfully created port: 7735c4ad-3f5d-440a-85ff-92b1270df99a {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 628.470634] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff1a17ad-cba2-47f5-ba4c-04884d149305 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.478694] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd1ee7b8-9b8c-4c9b-a831-71581d00155d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.509727] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56d3e27a-5bd2-4bc4-abda-36b1d2396890 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.518442] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0355ac8d-144a-46ae-a4b8-81c43e3c18de {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.534379] env[63531]: DEBUG nova.compute.manager [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 56872ede-96f0-48ca-b289-86784f37527d] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 628.538404] env[63531]: DEBUG nova.compute.provider_tree [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 628.664655] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7c6592cc-a823-47f6-859c-f1dc14b97b0f tempest-ServersV294TestFqdnHostnames-518322248 tempest-ServersV294TestFqdnHostnames-518322248-project-member] Lock "d941a593-7047-436a-b95d-bcb4850829f5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 64.420s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 629.043273] env[63531]: DEBUG nova.scheduler.client.report [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 629.166943] env[63531]: DEBUG nova.compute.manager [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 629.265830] env[63531]: ERROR nova.compute.manager [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7735c4ad-3f5d-440a-85ff-92b1270df99a, please check neutron logs for more information. [ 629.265830] env[63531]: ERROR nova.compute.manager Traceback (most recent call last): [ 629.265830] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 629.265830] env[63531]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 629.265830] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 629.265830] env[63531]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 629.265830] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 629.265830] env[63531]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 629.265830] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 629.265830] env[63531]: ERROR nova.compute.manager self.force_reraise() [ 629.265830] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 629.265830] env[63531]: ERROR nova.compute.manager raise self.value [ 629.265830] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 629.265830] env[63531]: ERROR nova.compute.manager updated_port = self._update_port( [ 629.265830] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 629.265830] env[63531]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 629.266219] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 629.266219] env[63531]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 629.266219] env[63531]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7735c4ad-3f5d-440a-85ff-92b1270df99a, please check neutron logs for more information. [ 629.266219] env[63531]: ERROR nova.compute.manager [ 629.266219] env[63531]: Traceback (most recent call last): [ 629.266219] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 629.266219] env[63531]: listener.cb(fileno) [ 629.266219] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 629.266219] env[63531]: result = function(*args, **kwargs) [ 629.266219] env[63531]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 629.266219] env[63531]: return func(*args, **kwargs) [ 629.266219] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 629.266219] env[63531]: raise e [ 629.266219] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 629.266219] env[63531]: nwinfo = self.network_api.allocate_for_instance( [ 629.266219] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 629.266219] env[63531]: created_port_ids = self._update_ports_for_instance( [ 629.266219] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 629.266219] env[63531]: with excutils.save_and_reraise_exception(): [ 629.266219] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 629.266219] env[63531]: self.force_reraise() [ 629.266219] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 629.266219] env[63531]: raise self.value [ 629.266219] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 629.266219] env[63531]: updated_port = self._update_port( [ 629.266219] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 629.266219] env[63531]: _ensure_no_port_binding_failure(port) [ 629.266219] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 629.266219] env[63531]: raise exception.PortBindingFailed(port_id=port['id']) [ 629.266817] env[63531]: nova.exception.PortBindingFailed: Binding failed for port 7735c4ad-3f5d-440a-85ff-92b1270df99a, please check neutron logs for more information. [ 629.266817] env[63531]: Removing descriptor: 17 [ 629.321240] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Acquiring lock "e1693594-d8bb-4002-b71c-7013efb81d16" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 629.321836] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Lock "e1693594-d8bb-4002-b71c-7013efb81d16" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 629.341679] env[63531]: DEBUG nova.compute.manager [req-4c0b2fe6-f6f5-4d1d-953f-79153e5d1cd3 req-7468c160-1e8d-4e86-8add-e9aa4497fdfd service nova] [instance: 56872ede-96f0-48ca-b289-86784f37527d] Received event network-changed-7735c4ad-3f5d-440a-85ff-92b1270df99a {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 629.343338] env[63531]: DEBUG nova.compute.manager [req-4c0b2fe6-f6f5-4d1d-953f-79153e5d1cd3 req-7468c160-1e8d-4e86-8add-e9aa4497fdfd service nova] [instance: 56872ede-96f0-48ca-b289-86784f37527d] Refreshing instance network info cache due to event network-changed-7735c4ad-3f5d-440a-85ff-92b1270df99a. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 629.343338] env[63531]: DEBUG oslo_concurrency.lockutils [req-4c0b2fe6-f6f5-4d1d-953f-79153e5d1cd3 req-7468c160-1e8d-4e86-8add-e9aa4497fdfd service nova] Acquiring lock "refresh_cache-56872ede-96f0-48ca-b289-86784f37527d" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 629.343338] env[63531]: DEBUG oslo_concurrency.lockutils [req-4c0b2fe6-f6f5-4d1d-953f-79153e5d1cd3 req-7468c160-1e8d-4e86-8add-e9aa4497fdfd service nova] Acquired lock "refresh_cache-56872ede-96f0-48ca-b289-86784f37527d" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 629.343338] env[63531]: DEBUG nova.network.neutron [req-4c0b2fe6-f6f5-4d1d-953f-79153e5d1cd3 req-7468c160-1e8d-4e86-8add-e9aa4497fdfd service nova] [instance: 56872ede-96f0-48ca-b289-86784f37527d] Refreshing network info cache for port 7735c4ad-3f5d-440a-85ff-92b1270df99a {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 629.552251] env[63531]: DEBUG nova.compute.manager [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 56872ede-96f0-48ca-b289-86784f37527d] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 629.554930] env[63531]: DEBUG oslo_concurrency.lockutils [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.038s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 629.555572] env[63531]: ERROR nova.compute.manager [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2c3c54ee-74be-4a1e-bfa8-a24d3e562f20, please check neutron logs for more information. [ 629.555572] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] Traceback (most recent call last): [ 629.555572] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 629.555572] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] self.driver.spawn(context, instance, image_meta, [ 629.555572] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 629.555572] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] self._vmops.spawn(context, instance, image_meta, injected_files, [ 629.555572] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 629.555572] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] vm_ref = self.build_virtual_machine(instance, [ 629.555572] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 629.555572] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] vif_infos = vmwarevif.get_vif_info(self._session, [ 629.555572] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 629.555860] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] for vif in network_info: [ 629.555860] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 629.555860] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] return self._sync_wrapper(fn, *args, **kwargs) [ 629.555860] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 629.555860] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] self.wait() [ 629.555860] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 629.555860] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] self[:] = self._gt.wait() [ 629.555860] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 629.555860] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] return self._exit_event.wait() [ 629.555860] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 629.555860] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] result = hub.switch() [ 629.555860] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 629.555860] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] return self.greenlet.switch() [ 629.556219] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 629.556219] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] result = function(*args, **kwargs) [ 629.556219] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 629.556219] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] return func(*args, **kwargs) [ 629.556219] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 629.556219] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] raise e [ 629.556219] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 629.556219] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] nwinfo = self.network_api.allocate_for_instance( [ 629.556219] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 629.556219] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] created_port_ids = self._update_ports_for_instance( [ 629.556219] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 629.556219] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] with excutils.save_and_reraise_exception(): [ 629.556219] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 629.556690] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] self.force_reraise() [ 629.556690] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 629.556690] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] raise self.value [ 629.556690] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 629.556690] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] updated_port = self._update_port( [ 629.556690] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 629.556690] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] _ensure_no_port_binding_failure(port) [ 629.556690] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 629.556690] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] raise exception.PortBindingFailed(port_id=port['id']) [ 629.556690] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] nova.exception.PortBindingFailed: Binding failed for port 2c3c54ee-74be-4a1e-bfa8-a24d3e562f20, please check neutron logs for more information. [ 629.556690] env[63531]: ERROR nova.compute.manager [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] [ 629.556958] env[63531]: DEBUG nova.compute.utils [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] Binding failed for port 2c3c54ee-74be-4a1e-bfa8-a24d3e562f20, please check neutron logs for more information. {{(pid=63531) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 629.557363] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 22.062s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 629.563725] env[63531]: DEBUG nova.compute.manager [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] Build of instance 2b5e3059-2772-496b-95c7-a083a37ef155 was re-scheduled: Binding failed for port 2c3c54ee-74be-4a1e-bfa8-a24d3e562f20, please check neutron logs for more information. {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 629.563725] env[63531]: DEBUG nova.compute.manager [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] Unplugging VIFs for instance {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 629.563725] env[63531]: DEBUG oslo_concurrency.lockutils [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Acquiring lock "refresh_cache-2b5e3059-2772-496b-95c7-a083a37ef155" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 629.563725] env[63531]: DEBUG oslo_concurrency.lockutils [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Acquired lock "refresh_cache-2b5e3059-2772-496b-95c7-a083a37ef155" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 629.564061] env[63531]: DEBUG nova.network.neutron [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 629.579879] env[63531]: DEBUG nova.virt.hardware [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 629.580124] env[63531]: DEBUG nova.virt.hardware [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 629.580280] env[63531]: DEBUG nova.virt.hardware [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 629.580463] env[63531]: DEBUG nova.virt.hardware [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 629.580609] env[63531]: DEBUG nova.virt.hardware [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 629.580755] env[63531]: DEBUG nova.virt.hardware [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 629.580959] env[63531]: DEBUG nova.virt.hardware [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 629.581139] env[63531]: DEBUG nova.virt.hardware [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 629.581302] env[63531]: DEBUG nova.virt.hardware [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 629.581507] env[63531]: DEBUG nova.virt.hardware [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 629.581719] env[63531]: DEBUG nova.virt.hardware [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 629.582915] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f30b6a8e-fa44-4b3a-870c-fd8689f69f0c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.595200] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0938820a-6090-4f1a-ac9e-17d001de6d72 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.611020] env[63531]: ERROR nova.compute.manager [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 56872ede-96f0-48ca-b289-86784f37527d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7735c4ad-3f5d-440a-85ff-92b1270df99a, please check neutron logs for more information. [ 629.611020] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] Traceback (most recent call last): [ 629.611020] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 629.611020] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] yield resources [ 629.611020] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 629.611020] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] self.driver.spawn(context, instance, image_meta, [ 629.611020] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 629.611020] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 629.611020] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 629.611020] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] vm_ref = self.build_virtual_machine(instance, [ 629.611020] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 629.611378] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] vif_infos = vmwarevif.get_vif_info(self._session, [ 629.611378] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 629.611378] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] for vif in network_info: [ 629.611378] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 629.611378] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] return self._sync_wrapper(fn, *args, **kwargs) [ 629.611378] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 629.611378] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] self.wait() [ 629.611378] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 629.611378] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] self[:] = self._gt.wait() [ 629.611378] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 629.611378] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] return self._exit_event.wait() [ 629.611378] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 629.611378] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] current.throw(*self._exc) [ 629.611676] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 629.611676] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] result = function(*args, **kwargs) [ 629.611676] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 629.611676] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] return func(*args, **kwargs) [ 629.611676] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 629.611676] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] raise e [ 629.611676] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 629.611676] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] nwinfo = self.network_api.allocate_for_instance( [ 629.611676] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 629.611676] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] created_port_ids = self._update_ports_for_instance( [ 629.611676] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 629.611676] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] with excutils.save_and_reraise_exception(): [ 629.611676] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 629.611949] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] self.force_reraise() [ 629.611949] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 629.611949] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] raise self.value [ 629.611949] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 629.611949] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] updated_port = self._update_port( [ 629.611949] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 629.611949] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] _ensure_no_port_binding_failure(port) [ 629.611949] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 629.611949] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] raise exception.PortBindingFailed(port_id=port['id']) [ 629.611949] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] nova.exception.PortBindingFailed: Binding failed for port 7735c4ad-3f5d-440a-85ff-92b1270df99a, please check neutron logs for more information. [ 629.611949] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] [ 629.611949] env[63531]: INFO nova.compute.manager [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 56872ede-96f0-48ca-b289-86784f37527d] Terminating instance [ 629.613092] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Acquiring lock "refresh_cache-56872ede-96f0-48ca-b289-86784f37527d" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 629.695023] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 629.864820] env[63531]: DEBUG nova.network.neutron [req-4c0b2fe6-f6f5-4d1d-953f-79153e5d1cd3 req-7468c160-1e8d-4e86-8add-e9aa4497fdfd service nova] [instance: 56872ede-96f0-48ca-b289-86784f37527d] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 629.996265] env[63531]: DEBUG nova.network.neutron [req-4c0b2fe6-f6f5-4d1d-953f-79153e5d1cd3 req-7468c160-1e8d-4e86-8add-e9aa4497fdfd service nova] [instance: 56872ede-96f0-48ca-b289-86784f37527d] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.091230] env[63531]: DEBUG nova.network.neutron [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 630.205596] env[63531]: DEBUG nova.network.neutron [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.501515] env[63531]: DEBUG oslo_concurrency.lockutils [req-4c0b2fe6-f6f5-4d1d-953f-79153e5d1cd3 req-7468c160-1e8d-4e86-8add-e9aa4497fdfd service nova] Releasing lock "refresh_cache-56872ede-96f0-48ca-b289-86784f37527d" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 630.501912] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Acquired lock "refresh_cache-56872ede-96f0-48ca-b289-86784f37527d" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 630.502105] env[63531]: DEBUG nova.network.neutron [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 56872ede-96f0-48ca-b289-86784f37527d] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 630.518932] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ad56ca6-c7b4-4f9e-8166-d814e542838f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.526684] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d25dca7e-5f0c-474f-bcaf-d2221ee17cc3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.560361] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59cda69e-99e0-4d03-b4cd-5bb3b8fa733f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.568342] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16bfe5e2-0e57-43b8-ac20-a2b1e524c436 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.581990] env[63531]: DEBUG nova.compute.provider_tree [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 630.707941] env[63531]: DEBUG oslo_concurrency.lockutils [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Releasing lock "refresh_cache-2b5e3059-2772-496b-95c7-a083a37ef155" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 630.708204] env[63531]: DEBUG nova.compute.manager [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 630.708389] env[63531]: DEBUG nova.compute.manager [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 630.708566] env[63531]: DEBUG nova.network.neutron [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 630.729695] env[63531]: DEBUG nova.network.neutron [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 631.029196] env[63531]: DEBUG nova.network.neutron [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 56872ede-96f0-48ca-b289-86784f37527d] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 631.085101] env[63531]: DEBUG nova.scheduler.client.report [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 631.162344] env[63531]: DEBUG nova.network.neutron [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 56872ede-96f0-48ca-b289-86784f37527d] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.232758] env[63531]: DEBUG nova.network.neutron [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.392093] env[63531]: DEBUG nova.compute.manager [req-4262a9b8-f372-4da9-ae20-5595e705bbc3 req-43193652-8bea-4246-9053-11cf4dcbb708 service nova] [instance: 56872ede-96f0-48ca-b289-86784f37527d] Received event network-vif-deleted-7735c4ad-3f5d-440a-85ff-92b1270df99a {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 631.592135] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.035s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 631.592817] env[63531]: ERROR nova.compute.manager [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3670ef9d-bf2d-4638-a691-74ed5d418a05, please check neutron logs for more information. [ 631.592817] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] Traceback (most recent call last): [ 631.592817] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 631.592817] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] self.driver.spawn(context, instance, image_meta, [ 631.592817] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 631.592817] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 631.592817] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 631.592817] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] vm_ref = self.build_virtual_machine(instance, [ 631.592817] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 631.592817] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] vif_infos = vmwarevif.get_vif_info(self._session, [ 631.592817] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 631.593081] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] for vif in network_info: [ 631.593081] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 631.593081] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] return self._sync_wrapper(fn, *args, **kwargs) [ 631.593081] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 631.593081] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] self.wait() [ 631.593081] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 631.593081] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] self[:] = self._gt.wait() [ 631.593081] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 631.593081] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] return self._exit_event.wait() [ 631.593081] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 631.593081] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] result = hub.switch() [ 631.593081] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 631.593081] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] return self.greenlet.switch() [ 631.593372] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 631.593372] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] result = function(*args, **kwargs) [ 631.593372] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 631.593372] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] return func(*args, **kwargs) [ 631.593372] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 631.593372] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] raise e [ 631.593372] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 631.593372] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] nwinfo = self.network_api.allocate_for_instance( [ 631.593372] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 631.593372] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] created_port_ids = self._update_ports_for_instance( [ 631.593372] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 631.593372] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] with excutils.save_and_reraise_exception(): [ 631.593372] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 631.593644] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] self.force_reraise() [ 631.593644] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 631.593644] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] raise self.value [ 631.593644] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 631.593644] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] updated_port = self._update_port( [ 631.593644] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 631.593644] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] _ensure_no_port_binding_failure(port) [ 631.593644] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 631.593644] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] raise exception.PortBindingFailed(port_id=port['id']) [ 631.593644] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] nova.exception.PortBindingFailed: Binding failed for port 3670ef9d-bf2d-4638-a691-74ed5d418a05, please check neutron logs for more information. [ 631.593644] env[63531]: ERROR nova.compute.manager [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] [ 631.593885] env[63531]: DEBUG nova.compute.utils [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] Binding failed for port 3670ef9d-bf2d-4638-a691-74ed5d418a05, please check neutron logs for more information. {{(pid=63531) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 631.596056] env[63531]: DEBUG oslo_concurrency.lockutils [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 21.201s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 631.603180] env[63531]: DEBUG nova.compute.manager [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] Build of instance 01a1b3e0-4ad9-4350-b59e-499f1b3412a5 was re-scheduled: Binding failed for port 3670ef9d-bf2d-4638-a691-74ed5d418a05, please check neutron logs for more information. {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 631.603620] env[63531]: DEBUG nova.compute.manager [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] Unplugging VIFs for instance {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 631.607996] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] Acquiring lock "refresh_cache-01a1b3e0-4ad9-4350-b59e-499f1b3412a5" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 631.607996] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] Acquired lock "refresh_cache-01a1b3e0-4ad9-4350-b59e-499f1b3412a5" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 631.607996] env[63531]: DEBUG nova.network.neutron [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 631.663611] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Releasing lock "refresh_cache-56872ede-96f0-48ca-b289-86784f37527d" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 631.664084] env[63531]: DEBUG nova.compute.manager [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 56872ede-96f0-48ca-b289-86784f37527d] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 631.664318] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 56872ede-96f0-48ca-b289-86784f37527d] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 631.664615] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-20a77a3c-0bd6-454a-97f1-23a0552b61e9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.673493] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf144136-426e-44bb-8f4f-081212f03ef5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.696852] env[63531]: WARNING nova.virt.vmwareapi.vmops [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 56872ede-96f0-48ca-b289-86784f37527d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 56872ede-96f0-48ca-b289-86784f37527d could not be found. [ 631.697095] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 56872ede-96f0-48ca-b289-86784f37527d] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 631.697299] env[63531]: INFO nova.compute.manager [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 56872ede-96f0-48ca-b289-86784f37527d] Took 0.03 seconds to destroy the instance on the hypervisor. [ 631.697573] env[63531]: DEBUG oslo.service.loopingcall [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 631.697803] env[63531]: DEBUG nova.compute.manager [-] [instance: 56872ede-96f0-48ca-b289-86784f37527d] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 631.697897] env[63531]: DEBUG nova.network.neutron [-] [instance: 56872ede-96f0-48ca-b289-86784f37527d] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 631.717494] env[63531]: DEBUG nova.network.neutron [-] [instance: 56872ede-96f0-48ca-b289-86784f37527d] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 631.735065] env[63531]: INFO nova.compute.manager [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 2b5e3059-2772-496b-95c7-a083a37ef155] Took 1.03 seconds to deallocate network for instance. [ 632.136960] env[63531]: DEBUG nova.network.neutron [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 632.220318] env[63531]: DEBUG nova.network.neutron [-] [instance: 56872ede-96f0-48ca-b289-86784f37527d] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.234131] env[63531]: DEBUG nova.network.neutron [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.522034] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0857ac8d-4974-4902-914d-83cb0d62274a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.528937] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a515aae1-8e68-47a4-b793-d503169eb5cb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.558684] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f94df05-4844-48da-a5d1-d5084635f18c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.566117] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bff1740e-b49e-4c4f-bb89-a38c8bf84783 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.579447] env[63531]: DEBUG nova.compute.provider_tree [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 632.722742] env[63531]: INFO nova.compute.manager [-] [instance: 56872ede-96f0-48ca-b289-86784f37527d] Took 1.02 seconds to deallocate network for instance. [ 632.725374] env[63531]: DEBUG nova.compute.claims [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 56872ede-96f0-48ca-b289-86784f37527d] Aborting claim: {{(pid=63531) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 632.725569] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 632.736240] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] Releasing lock "refresh_cache-01a1b3e0-4ad9-4350-b59e-499f1b3412a5" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 632.736897] env[63531]: DEBUG nova.compute.manager [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 632.736897] env[63531]: DEBUG nova.compute.manager [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 632.737220] env[63531]: DEBUG nova.network.neutron [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 632.753865] env[63531]: DEBUG nova.network.neutron [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 632.769768] env[63531]: INFO nova.scheduler.client.report [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Deleted allocations for instance 2b5e3059-2772-496b-95c7-a083a37ef155 [ 633.082971] env[63531]: DEBUG nova.scheduler.client.report [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 633.256130] env[63531]: DEBUG nova.network.neutron [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.281357] env[63531]: DEBUG oslo_concurrency.lockutils [None req-80f20955-1448-4587-afcc-aa6d8c48a2b5 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Lock "2b5e3059-2772-496b-95c7-a083a37ef155" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 68.315s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 633.590026] env[63531]: DEBUG oslo_concurrency.lockutils [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.992s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 633.590026] env[63531]: ERROR nova.compute.manager [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c30c9653-38f5-42cc-bda4-656ad1a333fc, please check neutron logs for more information. [ 633.590026] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] Traceback (most recent call last): [ 633.590026] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 633.590026] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] self.driver.spawn(context, instance, image_meta, [ 633.590026] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 633.590026] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] self._vmops.spawn(context, instance, image_meta, injected_files, [ 633.590026] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 633.590026] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] vm_ref = self.build_virtual_machine(instance, [ 633.590412] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 633.590412] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] vif_infos = vmwarevif.get_vif_info(self._session, [ 633.590412] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 633.590412] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] for vif in network_info: [ 633.590412] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 633.590412] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] return self._sync_wrapper(fn, *args, **kwargs) [ 633.590412] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 633.590412] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] self.wait() [ 633.590412] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 633.590412] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] self[:] = self._gt.wait() [ 633.590412] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 633.590412] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] return self._exit_event.wait() [ 633.590412] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 633.590700] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] result = hub.switch() [ 633.590700] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 633.590700] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] return self.greenlet.switch() [ 633.590700] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 633.590700] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] result = function(*args, **kwargs) [ 633.590700] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 633.590700] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] return func(*args, **kwargs) [ 633.590700] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 633.590700] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] raise e [ 633.590700] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 633.590700] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] nwinfo = self.network_api.allocate_for_instance( [ 633.590700] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 633.590700] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] created_port_ids = self._update_ports_for_instance( [ 633.590983] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 633.590983] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] with excutils.save_and_reraise_exception(): [ 633.590983] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 633.590983] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] self.force_reraise() [ 633.590983] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 633.590983] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] raise self.value [ 633.590983] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 633.590983] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] updated_port = self._update_port( [ 633.590983] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 633.590983] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] _ensure_no_port_binding_failure(port) [ 633.590983] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 633.590983] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] raise exception.PortBindingFailed(port_id=port['id']) [ 633.591260] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] nova.exception.PortBindingFailed: Binding failed for port c30c9653-38f5-42cc-bda4-656ad1a333fc, please check neutron logs for more information. [ 633.591260] env[63531]: ERROR nova.compute.manager [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] [ 633.591260] env[63531]: DEBUG nova.compute.utils [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] Binding failed for port c30c9653-38f5-42cc-bda4-656ad1a333fc, please check neutron logs for more information. {{(pid=63531) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 633.591567] env[63531]: DEBUG nova.compute.manager [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] Build of instance 8ac0012a-0855-4b73-a1e4-0a68a1336404 was re-scheduled: Binding failed for port c30c9653-38f5-42cc-bda4-656ad1a333fc, please check neutron logs for more information. {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 633.591928] env[63531]: DEBUG nova.compute.manager [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] Unplugging VIFs for instance {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 633.592165] env[63531]: DEBUG oslo_concurrency.lockutils [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] Acquiring lock "refresh_cache-8ac0012a-0855-4b73-a1e4-0a68a1336404" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 633.592303] env[63531]: DEBUG oslo_concurrency.lockutils [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] Acquired lock "refresh_cache-8ac0012a-0855-4b73-a1e4-0a68a1336404" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 633.592452] env[63531]: DEBUG nova.network.neutron [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 633.593873] env[63531]: DEBUG oslo_concurrency.lockutils [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.115s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 633.595354] env[63531]: INFO nova.compute.claims [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 633.761000] env[63531]: INFO nova.compute.manager [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] [instance: 01a1b3e0-4ad9-4350-b59e-499f1b3412a5] Took 1.02 seconds to deallocate network for instance. [ 633.785964] env[63531]: DEBUG nova.compute.manager [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 634.117218] env[63531]: DEBUG nova.network.neutron [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 634.259300] env[63531]: DEBUG nova.network.neutron [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.311120] env[63531]: DEBUG oslo_concurrency.lockutils [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 634.762292] env[63531]: DEBUG oslo_concurrency.lockutils [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] Releasing lock "refresh_cache-8ac0012a-0855-4b73-a1e4-0a68a1336404" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 634.762390] env[63531]: DEBUG nova.compute.manager [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 634.762559] env[63531]: DEBUG nova.compute.manager [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 634.762733] env[63531]: DEBUG nova.network.neutron [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 634.786894] env[63531]: DEBUG nova.network.neutron [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 634.794713] env[63531]: INFO nova.scheduler.client.report [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] Deleted allocations for instance 01a1b3e0-4ad9-4350-b59e-499f1b3412a5 [ 635.063574] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c75cf2f5-b671-4123-9e76-201ab202f370 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.072122] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1faa625-9ee9-4e07-b323-a66ecc16142a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.103462] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e764b209-aa56-42de-a9f3-45312cd63334 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.110921] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47606a45-c0e0-4b1b-9c83-b05e4604a546 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.123818] env[63531]: DEBUG nova.compute.provider_tree [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 635.292629] env[63531]: DEBUG nova.network.neutron [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 635.306118] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5c45319d-578a-4f8c-8e26-e2814d0e93b4 tempest-ServersWithSpecificFlavorTestJSON-1643396811 tempest-ServersWithSpecificFlavorTestJSON-1643396811-project-member] Lock "01a1b3e0-4ad9-4350-b59e-499f1b3412a5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 69.730s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 635.632824] env[63531]: DEBUG nova.scheduler.client.report [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 635.795914] env[63531]: INFO nova.compute.manager [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] [instance: 8ac0012a-0855-4b73-a1e4-0a68a1336404] Took 1.03 seconds to deallocate network for instance. [ 635.808606] env[63531]: DEBUG nova.compute.manager [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 636.143492] env[63531]: DEBUG oslo_concurrency.lockutils [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.549s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 636.144042] env[63531]: DEBUG nova.compute.manager [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 636.147513] env[63531]: DEBUG oslo_concurrency.lockutils [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.347s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 636.339733] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 636.661145] env[63531]: DEBUG nova.compute.utils [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 636.666291] env[63531]: DEBUG nova.compute.manager [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 636.667185] env[63531]: DEBUG nova.network.neutron [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 636.724145] env[63531]: DEBUG nova.policy [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3ab0a7a9f21d4a43a97a730963a3d648', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '77f5f9a8168847a9adc931bd6bf3051e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 636.835764] env[63531]: INFO nova.scheduler.client.report [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] Deleted allocations for instance 8ac0012a-0855-4b73-a1e4-0a68a1336404 [ 637.039048] env[63531]: DEBUG nova.network.neutron [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] Successfully created port: 06b2989d-8fee-41e1-bb1c-931c524779fd {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 637.129026] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "29dd6de3-2765-408c-acc0-da47e5e0a977" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 637.129273] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "29dd6de3-2765-408c-acc0-da47e5e0a977" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 637.133155] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ec6e448-4da4-4594-b79b-7915f09ede4b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.142776] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dab9a5ea-0a5f-4ee3-bc5d-0ede4b6c56bf {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.178893] env[63531]: DEBUG nova.compute.manager [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 637.182865] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1c696b5-996a-4756-b359-aa5b53fabfb2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.191277] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30534373-9f73-4908-94cf-380d19ba615f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.207395] env[63531]: DEBUG nova.compute.provider_tree [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 637.345941] env[63531]: DEBUG oslo_concurrency.lockutils [None req-70f0c26f-cc72-4b25-a1b0-d7a4538a00f5 tempest-VolumesAssistedSnapshotsTest-1462793897 tempest-VolumesAssistedSnapshotsTest-1462793897-project-member] Lock "8ac0012a-0855-4b73-a1e4-0a68a1336404" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 71.611s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 637.710138] env[63531]: DEBUG nova.scheduler.client.report [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 637.828669] env[63531]: DEBUG nova.compute.manager [req-f5b757f6-c87a-46d5-beae-c6f6de368e4a req-7e03ceaf-1a2c-45c6-bd09-2f64d3e592e6 service nova] [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] Received event network-changed-06b2989d-8fee-41e1-bb1c-931c524779fd {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 637.828854] env[63531]: DEBUG nova.compute.manager [req-f5b757f6-c87a-46d5-beae-c6f6de368e4a req-7e03ceaf-1a2c-45c6-bd09-2f64d3e592e6 service nova] [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] Refreshing instance network info cache due to event network-changed-06b2989d-8fee-41e1-bb1c-931c524779fd. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 637.829254] env[63531]: DEBUG oslo_concurrency.lockutils [req-f5b757f6-c87a-46d5-beae-c6f6de368e4a req-7e03ceaf-1a2c-45c6-bd09-2f64d3e592e6 service nova] Acquiring lock "refresh_cache-046304ef-b46f-473f-aa5b-5932b5078c54" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 637.829449] env[63531]: DEBUG oslo_concurrency.lockutils [req-f5b757f6-c87a-46d5-beae-c6f6de368e4a req-7e03ceaf-1a2c-45c6-bd09-2f64d3e592e6 service nova] Acquired lock "refresh_cache-046304ef-b46f-473f-aa5b-5932b5078c54" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 637.829704] env[63531]: DEBUG nova.network.neutron [req-f5b757f6-c87a-46d5-beae-c6f6de368e4a req-7e03ceaf-1a2c-45c6-bd09-2f64d3e592e6 service nova] [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] Refreshing network info cache for port 06b2989d-8fee-41e1-bb1c-931c524779fd {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 637.848939] env[63531]: DEBUG nova.compute.manager [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 637.947908] env[63531]: ERROR nova.compute.manager [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 06b2989d-8fee-41e1-bb1c-931c524779fd, please check neutron logs for more information. [ 637.947908] env[63531]: ERROR nova.compute.manager Traceback (most recent call last): [ 637.947908] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 637.947908] env[63531]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 637.947908] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 637.947908] env[63531]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 637.947908] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 637.947908] env[63531]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 637.947908] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 637.947908] env[63531]: ERROR nova.compute.manager self.force_reraise() [ 637.947908] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 637.947908] env[63531]: ERROR nova.compute.manager raise self.value [ 637.947908] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 637.947908] env[63531]: ERROR nova.compute.manager updated_port = self._update_port( [ 637.947908] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 637.947908] env[63531]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 637.948348] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 637.948348] env[63531]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 637.948348] env[63531]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 06b2989d-8fee-41e1-bb1c-931c524779fd, please check neutron logs for more information. [ 637.948348] env[63531]: ERROR nova.compute.manager [ 637.948348] env[63531]: Traceback (most recent call last): [ 637.948348] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 637.948348] env[63531]: listener.cb(fileno) [ 637.948348] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 637.948348] env[63531]: result = function(*args, **kwargs) [ 637.948348] env[63531]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 637.948348] env[63531]: return func(*args, **kwargs) [ 637.948348] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 637.948348] env[63531]: raise e [ 637.948348] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 637.948348] env[63531]: nwinfo = self.network_api.allocate_for_instance( [ 637.948348] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 637.948348] env[63531]: created_port_ids = self._update_ports_for_instance( [ 637.948348] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 637.948348] env[63531]: with excutils.save_and_reraise_exception(): [ 637.948348] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 637.948348] env[63531]: self.force_reraise() [ 637.948348] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 637.948348] env[63531]: raise self.value [ 637.948348] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 637.948348] env[63531]: updated_port = self._update_port( [ 637.948348] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 637.948348] env[63531]: _ensure_no_port_binding_failure(port) [ 637.948348] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 637.948348] env[63531]: raise exception.PortBindingFailed(port_id=port['id']) [ 637.949143] env[63531]: nova.exception.PortBindingFailed: Binding failed for port 06b2989d-8fee-41e1-bb1c-931c524779fd, please check neutron logs for more information. [ 637.949143] env[63531]: Removing descriptor: 17 [ 638.191319] env[63531]: DEBUG nova.compute.manager [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 638.218112] env[63531]: DEBUG oslo_concurrency.lockutils [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.069s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 638.218112] env[63531]: ERROR nova.compute.manager [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c90c3150-57cf-43ab-9ea6-c88af929ea4c, please check neutron logs for more information. [ 638.218112] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] Traceback (most recent call last): [ 638.218112] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 638.218112] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] self.driver.spawn(context, instance, image_meta, [ 638.218112] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 638.218112] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 638.218112] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 638.218112] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] vm_ref = self.build_virtual_machine(instance, [ 638.218379] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 638.218379] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] vif_infos = vmwarevif.get_vif_info(self._session, [ 638.218379] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 638.218379] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] for vif in network_info: [ 638.218379] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 638.218379] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] return self._sync_wrapper(fn, *args, **kwargs) [ 638.218379] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 638.218379] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] self.wait() [ 638.218379] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 638.218379] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] self[:] = self._gt.wait() [ 638.218379] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 638.218379] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] return self._exit_event.wait() [ 638.218379] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 638.218675] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] result = hub.switch() [ 638.218675] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 638.218675] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] return self.greenlet.switch() [ 638.218675] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 638.218675] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] result = function(*args, **kwargs) [ 638.218675] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 638.218675] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] return func(*args, **kwargs) [ 638.218675] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 638.218675] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] raise e [ 638.218675] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 638.218675] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] nwinfo = self.network_api.allocate_for_instance( [ 638.218675] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 638.218675] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] created_port_ids = self._update_ports_for_instance( [ 638.218968] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 638.218968] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] with excutils.save_and_reraise_exception(): [ 638.218968] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 638.218968] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] self.force_reraise() [ 638.218968] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 638.218968] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] raise self.value [ 638.218968] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 638.218968] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] updated_port = self._update_port( [ 638.218968] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 638.218968] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] _ensure_no_port_binding_failure(port) [ 638.218968] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 638.218968] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] raise exception.PortBindingFailed(port_id=port['id']) [ 638.219322] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] nova.exception.PortBindingFailed: Binding failed for port c90c3150-57cf-43ab-9ea6-c88af929ea4c, please check neutron logs for more information. [ 638.219322] env[63531]: ERROR nova.compute.manager [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] [ 638.219322] env[63531]: DEBUG nova.compute.utils [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] Binding failed for port c90c3150-57cf-43ab-9ea6-c88af929ea4c, please check neutron logs for more information. {{(pid=63531) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 638.219322] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.022s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 638.223902] env[63531]: DEBUG nova.compute.manager [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] Build of instance 3a02946a-64e9-41b7-95d7-0b32766d735c was re-scheduled: Binding failed for port c90c3150-57cf-43ab-9ea6-c88af929ea4c, please check neutron logs for more information. {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 638.223902] env[63531]: DEBUG nova.compute.manager [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] Unplugging VIFs for instance {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 638.223902] env[63531]: DEBUG oslo_concurrency.lockutils [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Acquiring lock "refresh_cache-3a02946a-64e9-41b7-95d7-0b32766d735c" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 638.223902] env[63531]: DEBUG oslo_concurrency.lockutils [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Acquired lock "refresh_cache-3a02946a-64e9-41b7-95d7-0b32766d735c" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.225331] env[63531]: DEBUG nova.network.neutron [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 638.227864] env[63531]: DEBUG nova.virt.hardware [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 638.227864] env[63531]: DEBUG nova.virt.hardware [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 638.227864] env[63531]: DEBUG nova.virt.hardware [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 638.228088] env[63531]: DEBUG nova.virt.hardware [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 638.228088] env[63531]: DEBUG nova.virt.hardware [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 638.228088] env[63531]: DEBUG nova.virt.hardware [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 638.228088] env[63531]: DEBUG nova.virt.hardware [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 638.228088] env[63531]: DEBUG nova.virt.hardware [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 638.228298] env[63531]: DEBUG nova.virt.hardware [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 638.228389] env[63531]: DEBUG nova.virt.hardware [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 638.228930] env[63531]: DEBUG nova.virt.hardware [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 638.229891] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e881883e-30a1-49c7-99df-73a0ca3ec1b6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.242032] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29661e9b-74a8-4103-83a5-7f2ec329b19f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.253729] env[63531]: ERROR nova.compute.manager [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 06b2989d-8fee-41e1-bb1c-931c524779fd, please check neutron logs for more information. [ 638.253729] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] Traceback (most recent call last): [ 638.253729] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 638.253729] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] yield resources [ 638.253729] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 638.253729] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] self.driver.spawn(context, instance, image_meta, [ 638.253729] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 638.253729] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] self._vmops.spawn(context, instance, image_meta, injected_files, [ 638.253729] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 638.253729] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] vm_ref = self.build_virtual_machine(instance, [ 638.253729] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 638.254038] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] vif_infos = vmwarevif.get_vif_info(self._session, [ 638.254038] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 638.254038] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] for vif in network_info: [ 638.254038] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 638.254038] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] return self._sync_wrapper(fn, *args, **kwargs) [ 638.254038] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 638.254038] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] self.wait() [ 638.254038] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 638.254038] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] self[:] = self._gt.wait() [ 638.254038] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 638.254038] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] return self._exit_event.wait() [ 638.254038] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 638.254038] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] current.throw(*self._exc) [ 638.254330] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 638.254330] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] result = function(*args, **kwargs) [ 638.254330] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 638.254330] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] return func(*args, **kwargs) [ 638.254330] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 638.254330] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] raise e [ 638.254330] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 638.254330] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] nwinfo = self.network_api.allocate_for_instance( [ 638.254330] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 638.254330] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] created_port_ids = self._update_ports_for_instance( [ 638.254330] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 638.254330] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] with excutils.save_and_reraise_exception(): [ 638.254330] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 638.254616] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] self.force_reraise() [ 638.254616] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 638.254616] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] raise self.value [ 638.254616] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 638.254616] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] updated_port = self._update_port( [ 638.254616] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 638.254616] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] _ensure_no_port_binding_failure(port) [ 638.254616] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 638.254616] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] raise exception.PortBindingFailed(port_id=port['id']) [ 638.254616] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] nova.exception.PortBindingFailed: Binding failed for port 06b2989d-8fee-41e1-bb1c-931c524779fd, please check neutron logs for more information. [ 638.254616] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] [ 638.255239] env[63531]: INFO nova.compute.manager [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] Terminating instance [ 638.259112] env[63531]: DEBUG oslo_concurrency.lockutils [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] Acquiring lock "refresh_cache-046304ef-b46f-473f-aa5b-5932b5078c54" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 638.366578] env[63531]: DEBUG nova.network.neutron [req-f5b757f6-c87a-46d5-beae-c6f6de368e4a req-7e03ceaf-1a2c-45c6-bd09-2f64d3e592e6 service nova] [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 638.369970] env[63531]: DEBUG oslo_concurrency.lockutils [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 638.482938] env[63531]: DEBUG nova.network.neutron [req-f5b757f6-c87a-46d5-beae-c6f6de368e4a req-7e03ceaf-1a2c-45c6-bd09-2f64d3e592e6 service nova] [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.774986] env[63531]: DEBUG nova.network.neutron [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 638.899062] env[63531]: DEBUG nova.network.neutron [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.986664] env[63531]: DEBUG oslo_concurrency.lockutils [req-f5b757f6-c87a-46d5-beae-c6f6de368e4a req-7e03ceaf-1a2c-45c6-bd09-2f64d3e592e6 service nova] Releasing lock "refresh_cache-046304ef-b46f-473f-aa5b-5932b5078c54" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 638.987203] env[63531]: DEBUG oslo_concurrency.lockutils [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] Acquired lock "refresh_cache-046304ef-b46f-473f-aa5b-5932b5078c54" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.987392] env[63531]: DEBUG nova.network.neutron [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 639.136987] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3468fdd-9758-4a27-ac8e-44cf45258204 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.144867] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f2820ba-2b27-4c62-b652-45d78aa41732 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.788468] env[63531]: DEBUG oslo_concurrency.lockutils [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Releasing lock "refresh_cache-3a02946a-64e9-41b7-95d7-0b32766d735c" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 639.788468] env[63531]: DEBUG nova.compute.manager [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 639.788468] env[63531]: DEBUG nova.compute.manager [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 639.788468] env[63531]: DEBUG nova.network.neutron [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 639.792100] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef38c31b-5225-4cd4-baa4-4c6f1fa1dd9d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.800068] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-194b3009-0196-44b0-b1e6-9f7d7ddfb858 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.813675] env[63531]: DEBUG nova.compute.provider_tree [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 639.818146] env[63531]: DEBUG nova.network.neutron [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 639.825924] env[63531]: DEBUG nova.network.neutron [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 639.922716] env[63531]: DEBUG nova.compute.manager [req-5eea7e37-f696-4c95-bc39-e080fb417a6f req-9e8a473c-8f85-4a12-8c62-d13ce99d4d6c service nova] [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] Received event network-vif-deleted-06b2989d-8fee-41e1-bb1c-931c524779fd {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 639.961038] env[63531]: DEBUG nova.network.neutron [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.317305] env[63531]: DEBUG nova.scheduler.client.report [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 640.323610] env[63531]: DEBUG nova.network.neutron [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.467021] env[63531]: DEBUG oslo_concurrency.lockutils [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] Releasing lock "refresh_cache-046304ef-b46f-473f-aa5b-5932b5078c54" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 640.467021] env[63531]: DEBUG nova.compute.manager [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 640.467021] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 640.467021] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7c57de52-5e18-4623-b395-f5b582ca0d1f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.476236] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e15284a-4a3b-4976-b16e-85e4700b4652 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.500680] env[63531]: WARNING nova.virt.vmwareapi.vmops [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 046304ef-b46f-473f-aa5b-5932b5078c54 could not be found. [ 640.500680] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 640.500680] env[63531]: INFO nova.compute.manager [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] Took 0.04 seconds to destroy the instance on the hypervisor. [ 640.500813] env[63531]: DEBUG oslo.service.loopingcall [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 640.501229] env[63531]: DEBUG nova.compute.manager [-] [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 640.501229] env[63531]: DEBUG nova.network.neutron [-] [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 640.518357] env[63531]: DEBUG nova.network.neutron [-] [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 640.828138] env[63531]: INFO nova.compute.manager [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 3a02946a-64e9-41b7-95d7-0b32766d735c] Took 1.04 seconds to deallocate network for instance. [ 640.832475] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.614s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 640.833151] env[63531]: ERROR nova.compute.manager [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 62dcd413-b725-4384-ba5a-9ccaffb40cd3, please check neutron logs for more information. [ 640.833151] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] Traceback (most recent call last): [ 640.833151] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 640.833151] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] self.driver.spawn(context, instance, image_meta, [ 640.833151] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 640.833151] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 640.833151] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 640.833151] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] vm_ref = self.build_virtual_machine(instance, [ 640.833151] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 640.833151] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] vif_infos = vmwarevif.get_vif_info(self._session, [ 640.833151] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 640.833511] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] for vif in network_info: [ 640.833511] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 640.833511] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] return self._sync_wrapper(fn, *args, **kwargs) [ 640.833511] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 640.833511] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] self.wait() [ 640.833511] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 640.833511] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] self[:] = self._gt.wait() [ 640.833511] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 640.833511] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] return self._exit_event.wait() [ 640.833511] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 640.833511] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] result = hub.switch() [ 640.833511] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 640.833511] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] return self.greenlet.switch() [ 640.834826] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 640.834826] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] result = function(*args, **kwargs) [ 640.834826] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 640.834826] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] return func(*args, **kwargs) [ 640.834826] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 640.834826] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] raise e [ 640.834826] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 640.834826] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] nwinfo = self.network_api.allocate_for_instance( [ 640.834826] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 640.834826] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] created_port_ids = self._update_ports_for_instance( [ 640.834826] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 640.834826] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] with excutils.save_and_reraise_exception(): [ 640.834826] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 640.835133] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] self.force_reraise() [ 640.835133] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 640.835133] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] raise self.value [ 640.835133] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 640.835133] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] updated_port = self._update_port( [ 640.835133] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 640.835133] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] _ensure_no_port_binding_failure(port) [ 640.835133] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 640.835133] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] raise exception.PortBindingFailed(port_id=port['id']) [ 640.835133] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] nova.exception.PortBindingFailed: Binding failed for port 62dcd413-b725-4384-ba5a-9ccaffb40cd3, please check neutron logs for more information. [ 640.835133] env[63531]: ERROR nova.compute.manager [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] [ 640.835597] env[63531]: DEBUG nova.compute.utils [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] Binding failed for port 62dcd413-b725-4384-ba5a-9ccaffb40cd3, please check neutron logs for more information. {{(pid=63531) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 640.835597] env[63531]: DEBUG oslo_concurrency.lockutils [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.283s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 640.838178] env[63531]: DEBUG nova.compute.manager [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] Build of instance fe881dbe-6d44-4e31-a537-4aa9dfe44ff4 was re-scheduled: Binding failed for port 62dcd413-b725-4384-ba5a-9ccaffb40cd3, please check neutron logs for more information. {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 640.838178] env[63531]: DEBUG nova.compute.manager [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] Unplugging VIFs for instance {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 640.838976] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] Acquiring lock "refresh_cache-fe881dbe-6d44-4e31-a537-4aa9dfe44ff4" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 640.838976] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] Acquired lock "refresh_cache-fe881dbe-6d44-4e31-a537-4aa9dfe44ff4" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 640.838976] env[63531]: DEBUG nova.network.neutron [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 641.020538] env[63531]: DEBUG nova.network.neutron [-] [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 641.361013] env[63531]: DEBUG nova.network.neutron [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 641.422975] env[63531]: DEBUG nova.network.neutron [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 641.524282] env[63531]: INFO nova.compute.manager [-] [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] Took 1.02 seconds to deallocate network for instance. [ 641.527212] env[63531]: DEBUG nova.compute.claims [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] Aborting claim: {{(pid=63531) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 641.527393] env[63531]: DEBUG oslo_concurrency.lockutils [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 641.803999] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f20831e-cfbe-47bc-a048-6a568ca1c0a8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.813275] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ffa19fc-9ee8-4997-8a87-4d32a95dac6f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.846698] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-772fb782-b6ae-4b15-bb26-602536a96b78 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.857524] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a994247f-300f-4b23-af0a-893263ed1dd4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.871742] env[63531]: DEBUG nova.compute.provider_tree [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 641.874365] env[63531]: INFO nova.scheduler.client.report [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Deleted allocations for instance 3a02946a-64e9-41b7-95d7-0b32766d735c [ 641.928019] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] Releasing lock "refresh_cache-fe881dbe-6d44-4e31-a537-4aa9dfe44ff4" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 641.928019] env[63531]: DEBUG nova.compute.manager [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 641.928019] env[63531]: DEBUG nova.compute.manager [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 641.928019] env[63531]: DEBUG nova.network.neutron [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 641.944089] env[63531]: DEBUG nova.network.neutron [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 642.380424] env[63531]: DEBUG nova.scheduler.client.report [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 642.386590] env[63531]: DEBUG oslo_concurrency.lockutils [None req-d230a962-fd08-4f01-a227-713aefb5e147 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Lock "3a02946a-64e9-41b7-95d7-0b32766d735c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 75.535s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 642.445442] env[63531]: DEBUG nova.network.neutron [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.886145] env[63531]: DEBUG oslo_concurrency.lockutils [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.051s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 642.887996] env[63531]: ERROR nova.compute.manager [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 584241fa-8b49-462d-8c0f-0accc2a17497, please check neutron logs for more information. [ 642.887996] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] Traceback (most recent call last): [ 642.887996] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 642.887996] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] self.driver.spawn(context, instance, image_meta, [ 642.887996] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 642.887996] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] self._vmops.spawn(context, instance, image_meta, injected_files, [ 642.887996] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 642.887996] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] vm_ref = self.build_virtual_machine(instance, [ 642.887996] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 642.887996] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] vif_infos = vmwarevif.get_vif_info(self._session, [ 642.887996] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 642.889257] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] for vif in network_info: [ 642.889257] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 642.889257] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] return self._sync_wrapper(fn, *args, **kwargs) [ 642.889257] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 642.889257] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] self.wait() [ 642.889257] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 642.889257] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] self[:] = self._gt.wait() [ 642.889257] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 642.889257] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] return self._exit_event.wait() [ 642.889257] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 642.889257] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] result = hub.switch() [ 642.889257] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 642.889257] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] return self.greenlet.switch() [ 642.889794] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 642.889794] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] result = function(*args, **kwargs) [ 642.889794] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 642.889794] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] return func(*args, **kwargs) [ 642.889794] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 642.889794] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] raise e [ 642.889794] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 642.889794] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] nwinfo = self.network_api.allocate_for_instance( [ 642.889794] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 642.889794] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] created_port_ids = self._update_ports_for_instance( [ 642.889794] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 642.889794] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] with excutils.save_and_reraise_exception(): [ 642.889794] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 642.890335] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] self.force_reraise() [ 642.890335] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 642.890335] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] raise self.value [ 642.890335] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 642.890335] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] updated_port = self._update_port( [ 642.890335] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 642.890335] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] _ensure_no_port_binding_failure(port) [ 642.890335] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 642.890335] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] raise exception.PortBindingFailed(port_id=port['id']) [ 642.890335] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] nova.exception.PortBindingFailed: Binding failed for port 584241fa-8b49-462d-8c0f-0accc2a17497, please check neutron logs for more information. [ 642.890335] env[63531]: ERROR nova.compute.manager [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] [ 642.890753] env[63531]: DEBUG nova.compute.utils [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] Binding failed for port 584241fa-8b49-462d-8c0f-0accc2a17497, please check neutron logs for more information. {{(pid=63531) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 642.890753] env[63531]: DEBUG nova.compute.manager [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 642.893596] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.184s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 642.895542] env[63531]: INFO nova.compute.claims [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 642.902016] env[63531]: DEBUG nova.compute.manager [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] Build of instance 3c513284-4860-46b2-8ba1-e185efe0db86 was re-scheduled: Binding failed for port 584241fa-8b49-462d-8c0f-0accc2a17497, please check neutron logs for more information. {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 642.902016] env[63531]: DEBUG nova.compute.manager [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] Unplugging VIFs for instance {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 642.902016] env[63531]: DEBUG oslo_concurrency.lockutils [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] Acquiring lock "refresh_cache-3c513284-4860-46b2-8ba1-e185efe0db86" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 642.902016] env[63531]: DEBUG oslo_concurrency.lockutils [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] Acquired lock "refresh_cache-3c513284-4860-46b2-8ba1-e185efe0db86" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 642.902216] env[63531]: DEBUG nova.network.neutron [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 642.948285] env[63531]: INFO nova.compute.manager [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] [instance: fe881dbe-6d44-4e31-a537-4aa9dfe44ff4] Took 1.02 seconds to deallocate network for instance. [ 643.428445] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 643.432866] env[63531]: DEBUG nova.network.neutron [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 643.533293] env[63531]: DEBUG nova.network.neutron [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.989454] env[63531]: INFO nova.scheduler.client.report [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] Deleted allocations for instance fe881dbe-6d44-4e31-a537-4aa9dfe44ff4 [ 644.035638] env[63531]: DEBUG oslo_concurrency.lockutils [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] Releasing lock "refresh_cache-3c513284-4860-46b2-8ba1-e185efe0db86" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 644.035948] env[63531]: DEBUG nova.compute.manager [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 644.036194] env[63531]: DEBUG nova.compute.manager [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 644.036423] env[63531]: DEBUG nova.network.neutron [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 644.065553] env[63531]: DEBUG nova.network.neutron [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 644.359604] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1eff8b8e-4c86-46de-b5bc-0339dd51bafd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.367337] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89e6a071-1007-41fd-84e0-502391935c80 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.400359] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b32fa9e-fbc2-4f55-bf5f-4a8158d163b5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.408175] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b5d7659-06b2-44ec-8437-08f4eebfbdce {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.422721] env[63531]: DEBUG nova.compute.provider_tree [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 644.498590] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1f8929e0-77cb-4545-b0f3-d902fef982c7 tempest-ServerDiagnosticsTest-1995507562 tempest-ServerDiagnosticsTest-1995507562-project-member] Lock "fe881dbe-6d44-4e31-a537-4aa9dfe44ff4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 77.455s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 644.568861] env[63531]: DEBUG nova.network.neutron [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 644.927325] env[63531]: DEBUG nova.scheduler.client.report [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 645.001215] env[63531]: DEBUG nova.compute.manager [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 645.075657] env[63531]: INFO nova.compute.manager [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] [instance: 3c513284-4860-46b2-8ba1-e185efe0db86] Took 1.04 seconds to deallocate network for instance. [ 645.433423] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.540s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 645.433958] env[63531]: DEBUG nova.compute.manager [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 645.440934] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.746s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 645.442346] env[63531]: INFO nova.compute.claims [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 645.535297] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 645.947200] env[63531]: DEBUG nova.compute.utils [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 645.948799] env[63531]: DEBUG nova.compute.manager [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 645.950910] env[63531]: DEBUG nova.network.neutron [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 646.016592] env[63531]: DEBUG nova.policy [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '191861b327214987bffa08e13e505179', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e91751496e084b83a489969825df9711', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 646.120023] env[63531]: INFO nova.scheduler.client.report [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] Deleted allocations for instance 3c513284-4860-46b2-8ba1-e185efe0db86 [ 646.390245] env[63531]: DEBUG nova.network.neutron [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] Successfully created port: d338839c-7a92-49b7-a6e9-e0abb9e43d33 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 646.454632] env[63531]: DEBUG nova.compute.manager [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 646.632120] env[63531]: DEBUG oslo_concurrency.lockutils [None req-658a1108-de47-4bf5-80f2-5593e919618e tempest-ImagesNegativeTestJSON-1253941886 tempest-ImagesNegativeTestJSON-1253941886-project-member] Lock "3c513284-4860-46b2-8ba1-e185efe0db86" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 79.121s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 646.889479] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbca8b83-38bb-4d05-b0cf-ad2a1491e991 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.897309] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f67a179e-d3c7-4b19-a080-4efbb417c3a7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.940240] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cd1a493-9e1a-4b4d-8142-be0c65cf4532 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.948630] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-650bfa90-b57c-42d8-8d20-18d4b1a39408 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.968910] env[63531]: DEBUG nova.compute.provider_tree [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 647.135278] env[63531]: DEBUG nova.compute.manager [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 647.470730] env[63531]: DEBUG nova.compute.manager [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 647.473605] env[63531]: DEBUG nova.scheduler.client.report [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 647.509550] env[63531]: DEBUG nova.virt.hardware [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 647.509814] env[63531]: DEBUG nova.virt.hardware [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 647.509912] env[63531]: DEBUG nova.virt.hardware [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 647.510564] env[63531]: DEBUG nova.virt.hardware [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 647.510775] env[63531]: DEBUG nova.virt.hardware [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 647.510892] env[63531]: DEBUG nova.virt.hardware [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 647.511991] env[63531]: DEBUG nova.virt.hardware [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 647.512069] env[63531]: DEBUG nova.virt.hardware [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 647.512246] env[63531]: DEBUG nova.virt.hardware [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 647.512403] env[63531]: DEBUG nova.virt.hardware [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 647.512604] env[63531]: DEBUG nova.virt.hardware [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 647.513760] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22ddd7df-9d1f-4fd7-b3f1-e59c3ffeb35a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.523771] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f35dd288-38a2-4d37-a306-63163f373079 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.660103] env[63531]: DEBUG oslo_concurrency.lockutils [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 647.696857] env[63531]: DEBUG nova.compute.manager [req-958c1485-96c8-410e-8336-156bd97191e5 req-272d46d9-b986-40f5-979c-46751f6008a0 service nova] [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] Received event network-changed-d338839c-7a92-49b7-a6e9-e0abb9e43d33 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 647.697049] env[63531]: DEBUG nova.compute.manager [req-958c1485-96c8-410e-8336-156bd97191e5 req-272d46d9-b986-40f5-979c-46751f6008a0 service nova] [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] Refreshing instance network info cache due to event network-changed-d338839c-7a92-49b7-a6e9-e0abb9e43d33. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 647.698369] env[63531]: DEBUG oslo_concurrency.lockutils [req-958c1485-96c8-410e-8336-156bd97191e5 req-272d46d9-b986-40f5-979c-46751f6008a0 service nova] Acquiring lock "refresh_cache-e1e72ba0-fc50-4776-991f-738c2d9c4ff3" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 647.698981] env[63531]: DEBUG oslo_concurrency.lockutils [req-958c1485-96c8-410e-8336-156bd97191e5 req-272d46d9-b986-40f5-979c-46751f6008a0 service nova] Acquired lock "refresh_cache-e1e72ba0-fc50-4776-991f-738c2d9c4ff3" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 647.699237] env[63531]: DEBUG nova.network.neutron [req-958c1485-96c8-410e-8336-156bd97191e5 req-272d46d9-b986-40f5-979c-46751f6008a0 service nova] [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] Refreshing network info cache for port d338839c-7a92-49b7-a6e9-e0abb9e43d33 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 647.788025] env[63531]: ERROR nova.compute.manager [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d338839c-7a92-49b7-a6e9-e0abb9e43d33, please check neutron logs for more information. [ 647.788025] env[63531]: ERROR nova.compute.manager Traceback (most recent call last): [ 647.788025] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 647.788025] env[63531]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 647.788025] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 647.788025] env[63531]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 647.788025] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 647.788025] env[63531]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 647.788025] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 647.788025] env[63531]: ERROR nova.compute.manager self.force_reraise() [ 647.788025] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 647.788025] env[63531]: ERROR nova.compute.manager raise self.value [ 647.788025] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 647.788025] env[63531]: ERROR nova.compute.manager updated_port = self._update_port( [ 647.788025] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 647.788025] env[63531]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 647.788513] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 647.788513] env[63531]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 647.788513] env[63531]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d338839c-7a92-49b7-a6e9-e0abb9e43d33, please check neutron logs for more information. [ 647.788513] env[63531]: ERROR nova.compute.manager [ 647.788513] env[63531]: Traceback (most recent call last): [ 647.788513] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 647.788513] env[63531]: listener.cb(fileno) [ 647.788513] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 647.788513] env[63531]: result = function(*args, **kwargs) [ 647.788513] env[63531]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 647.788513] env[63531]: return func(*args, **kwargs) [ 647.788513] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 647.788513] env[63531]: raise e [ 647.788513] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 647.788513] env[63531]: nwinfo = self.network_api.allocate_for_instance( [ 647.788513] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 647.788513] env[63531]: created_port_ids = self._update_ports_for_instance( [ 647.788513] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 647.788513] env[63531]: with excutils.save_and_reraise_exception(): [ 647.788513] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 647.788513] env[63531]: self.force_reraise() [ 647.788513] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 647.788513] env[63531]: raise self.value [ 647.788513] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 647.788513] env[63531]: updated_port = self._update_port( [ 647.788513] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 647.788513] env[63531]: _ensure_no_port_binding_failure(port) [ 647.788513] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 647.788513] env[63531]: raise exception.PortBindingFailed(port_id=port['id']) [ 647.789315] env[63531]: nova.exception.PortBindingFailed: Binding failed for port d338839c-7a92-49b7-a6e9-e0abb9e43d33, please check neutron logs for more information. [ 647.789315] env[63531]: Removing descriptor: 20 [ 647.789315] env[63531]: ERROR nova.compute.manager [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d338839c-7a92-49b7-a6e9-e0abb9e43d33, please check neutron logs for more information. [ 647.789315] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] Traceback (most recent call last): [ 647.789315] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 647.789315] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] yield resources [ 647.789315] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 647.789315] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] self.driver.spawn(context, instance, image_meta, [ 647.789315] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 647.789315] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 647.789315] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 647.789315] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] vm_ref = self.build_virtual_machine(instance, [ 647.789655] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 647.789655] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] vif_infos = vmwarevif.get_vif_info(self._session, [ 647.789655] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 647.789655] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] for vif in network_info: [ 647.789655] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 647.789655] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] return self._sync_wrapper(fn, *args, **kwargs) [ 647.789655] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 647.789655] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] self.wait() [ 647.789655] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 647.789655] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] self[:] = self._gt.wait() [ 647.789655] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 647.789655] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] return self._exit_event.wait() [ 647.789655] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 647.789980] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] result = hub.switch() [ 647.789980] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 647.789980] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] return self.greenlet.switch() [ 647.789980] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 647.789980] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] result = function(*args, **kwargs) [ 647.789980] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 647.789980] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] return func(*args, **kwargs) [ 647.789980] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 647.789980] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] raise e [ 647.789980] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 647.789980] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] nwinfo = self.network_api.allocate_for_instance( [ 647.789980] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 647.789980] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] created_port_ids = self._update_ports_for_instance( [ 647.790288] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 647.790288] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] with excutils.save_and_reraise_exception(): [ 647.790288] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 647.790288] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] self.force_reraise() [ 647.790288] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 647.790288] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] raise self.value [ 647.790288] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 647.790288] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] updated_port = self._update_port( [ 647.790288] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 647.790288] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] _ensure_no_port_binding_failure(port) [ 647.790288] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 647.790288] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] raise exception.PortBindingFailed(port_id=port['id']) [ 647.790617] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] nova.exception.PortBindingFailed: Binding failed for port d338839c-7a92-49b7-a6e9-e0abb9e43d33, please check neutron logs for more information. [ 647.790617] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] [ 647.790617] env[63531]: INFO nova.compute.manager [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] Terminating instance [ 647.792359] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] Acquiring lock "refresh_cache-e1e72ba0-fc50-4776-991f-738c2d9c4ff3" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 647.982462] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.541s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 647.982986] env[63531]: DEBUG nova.compute.manager [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 647.987773] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.262s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 648.227057] env[63531]: DEBUG nova.network.neutron [req-958c1485-96c8-410e-8336-156bd97191e5 req-272d46d9-b986-40f5-979c-46751f6008a0 service nova] [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 648.369756] env[63531]: DEBUG nova.network.neutron [req-958c1485-96c8-410e-8336-156bd97191e5 req-272d46d9-b986-40f5-979c-46751f6008a0 service nova] [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 648.489745] env[63531]: DEBUG nova.compute.utils [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 648.497212] env[63531]: DEBUG nova.compute.manager [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 648.497806] env[63531]: DEBUG nova.network.neutron [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 648.505726] env[63531]: DEBUG oslo_concurrency.lockutils [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Acquiring lock "dccb2335-4220-4570-861b-16ff2d8f5a85" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 648.505726] env[63531]: DEBUG oslo_concurrency.lockutils [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Lock "dccb2335-4220-4570-861b-16ff2d8f5a85" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 648.625032] env[63531]: DEBUG nova.policy [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '979d9701753b4d0082abe8b5ef217529', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4cb35716050e4e8c95948ca7913bf815', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 648.872866] env[63531]: DEBUG oslo_concurrency.lockutils [req-958c1485-96c8-410e-8336-156bd97191e5 req-272d46d9-b986-40f5-979c-46751f6008a0 service nova] Releasing lock "refresh_cache-e1e72ba0-fc50-4776-991f-738c2d9c4ff3" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 648.873328] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] Acquired lock "refresh_cache-e1e72ba0-fc50-4776-991f-738c2d9c4ff3" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 648.873810] env[63531]: DEBUG nova.network.neutron [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 648.937569] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ae8765d-e928-46ab-98a9-dc1f89511431 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.945702] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea82a3bb-f79e-4247-b032-a5b8409a920b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.979498] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cace0574-4be9-4be4-8f59-0bb91991adf2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.988329] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-436dd527-b901-4720-b096-26726b2c3660 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.001982] env[63531]: DEBUG nova.compute.provider_tree [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 649.005907] env[63531]: DEBUG nova.compute.manager [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 649.428274] env[63531]: DEBUG nova.network.neutron [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 649.501730] env[63531]: DEBUG nova.network.neutron [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] Successfully created port: b82524d5-0c39-474f-b157-e64fef4fa205 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 649.510761] env[63531]: DEBUG nova.scheduler.client.report [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 649.608188] env[63531]: DEBUG nova.network.neutron [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 649.860182] env[63531]: DEBUG nova.compute.manager [req-ce0e39a9-db4f-4e38-b3d5-59409ca0f8db req-0af1e421-a26a-43d8-a287-7244db7525aa service nova] [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] Received event network-vif-deleted-d338839c-7a92-49b7-a6e9-e0abb9e43d33 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 650.023946] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.033s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 650.023946] env[63531]: ERROR nova.compute.manager [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 56872ede-96f0-48ca-b289-86784f37527d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7735c4ad-3f5d-440a-85ff-92b1270df99a, please check neutron logs for more information. [ 650.023946] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] Traceback (most recent call last): [ 650.023946] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 650.023946] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] self.driver.spawn(context, instance, image_meta, [ 650.023946] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 650.023946] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 650.023946] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 650.023946] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] vm_ref = self.build_virtual_machine(instance, [ 650.024348] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 650.024348] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] vif_infos = vmwarevif.get_vif_info(self._session, [ 650.024348] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 650.024348] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] for vif in network_info: [ 650.024348] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 650.024348] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] return self._sync_wrapper(fn, *args, **kwargs) [ 650.024348] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 650.024348] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] self.wait() [ 650.024348] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 650.024348] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] self[:] = self._gt.wait() [ 650.024348] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 650.024348] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] return self._exit_event.wait() [ 650.024348] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 650.024666] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] current.throw(*self._exc) [ 650.024666] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 650.024666] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] result = function(*args, **kwargs) [ 650.024666] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 650.024666] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] return func(*args, **kwargs) [ 650.024666] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 650.024666] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] raise e [ 650.024666] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 650.024666] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] nwinfo = self.network_api.allocate_for_instance( [ 650.024666] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 650.024666] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] created_port_ids = self._update_ports_for_instance( [ 650.024666] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 650.024666] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] with excutils.save_and_reraise_exception(): [ 650.024983] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 650.024983] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] self.force_reraise() [ 650.024983] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 650.024983] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] raise self.value [ 650.024983] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 650.024983] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] updated_port = self._update_port( [ 650.024983] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 650.024983] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] _ensure_no_port_binding_failure(port) [ 650.024983] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 650.024983] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] raise exception.PortBindingFailed(port_id=port['id']) [ 650.024983] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] nova.exception.PortBindingFailed: Binding failed for port 7735c4ad-3f5d-440a-85ff-92b1270df99a, please check neutron logs for more information. [ 650.024983] env[63531]: ERROR nova.compute.manager [instance: 56872ede-96f0-48ca-b289-86784f37527d] [ 650.025333] env[63531]: DEBUG nova.compute.utils [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 56872ede-96f0-48ca-b289-86784f37527d] Binding failed for port 7735c4ad-3f5d-440a-85ff-92b1270df99a, please check neutron logs for more information. {{(pid=63531) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 650.025333] env[63531]: DEBUG oslo_concurrency.lockutils [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.712s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 650.025333] env[63531]: INFO nova.compute.claims [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 650.029486] env[63531]: DEBUG nova.compute.manager [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 650.034137] env[63531]: DEBUG nova.compute.manager [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 56872ede-96f0-48ca-b289-86784f37527d] Build of instance 56872ede-96f0-48ca-b289-86784f37527d was re-scheduled: Binding failed for port 7735c4ad-3f5d-440a-85ff-92b1270df99a, please check neutron logs for more information. {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 650.035218] env[63531]: DEBUG nova.compute.manager [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 56872ede-96f0-48ca-b289-86784f37527d] Unplugging VIFs for instance {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 650.035218] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Acquiring lock "refresh_cache-56872ede-96f0-48ca-b289-86784f37527d" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 650.035218] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Acquired lock "refresh_cache-56872ede-96f0-48ca-b289-86784f37527d" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 650.035218] env[63531]: DEBUG nova.network.neutron [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 56872ede-96f0-48ca-b289-86784f37527d] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 650.065945] env[63531]: DEBUG nova.virt.hardware [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 650.066240] env[63531]: DEBUG nova.virt.hardware [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 650.066421] env[63531]: DEBUG nova.virt.hardware [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 650.066614] env[63531]: DEBUG nova.virt.hardware [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 650.066761] env[63531]: DEBUG nova.virt.hardware [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 650.066904] env[63531]: DEBUG nova.virt.hardware [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 650.068316] env[63531]: DEBUG nova.virt.hardware [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 650.068316] env[63531]: DEBUG nova.virt.hardware [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 650.068467] env[63531]: DEBUG nova.virt.hardware [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 650.068641] env[63531]: DEBUG nova.virt.hardware [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 650.068816] env[63531]: DEBUG nova.virt.hardware [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 650.070787] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1fcce0f-8003-4d25-b8b7-955e4d6e2801 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.082328] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7208020b-8551-4c06-9323-af360fd432e5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.110719] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] Releasing lock "refresh_cache-e1e72ba0-fc50-4776-991f-738c2d9c4ff3" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 650.111846] env[63531]: DEBUG nova.compute.manager [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 650.111846] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 650.111846] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-598c2c1e-b7ac-4ca9-b075-de69cf1d404a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.120519] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bafcff0-027a-40bc-827b-eae0cf5f5772 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.144554] env[63531]: WARNING nova.virt.vmwareapi.vmops [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e1e72ba0-fc50-4776-991f-738c2d9c4ff3 could not be found. [ 650.144931] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 650.145228] env[63531]: INFO nova.compute.manager [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] Took 0.03 seconds to destroy the instance on the hypervisor. [ 650.145559] env[63531]: DEBUG oslo.service.loopingcall [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 650.145964] env[63531]: DEBUG nova.compute.manager [-] [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 650.146141] env[63531]: DEBUG nova.network.neutron [-] [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 650.175111] env[63531]: DEBUG nova.network.neutron [-] [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 650.197998] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 650.197998] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 650.575107] env[63531]: DEBUG nova.network.neutron [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 56872ede-96f0-48ca-b289-86784f37527d] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 650.677401] env[63531]: DEBUG nova.network.neutron [-] [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 650.707643] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 650.707818] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Starting heal instance info cache {{(pid=63531) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 650.707932] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Rebuilding the list of instances to heal {{(pid=63531) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 650.771566] env[63531]: DEBUG nova.network.neutron [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 56872ede-96f0-48ca-b289-86784f37527d] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 651.180588] env[63531]: INFO nova.compute.manager [-] [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] Took 1.03 seconds to deallocate network for instance. [ 651.188123] env[63531]: DEBUG nova.compute.claims [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] Aborting claim: {{(pid=63531) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 651.188123] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 651.216069] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] Skipping network cache update for instance because it is Building. {{(pid=63531) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 651.216914] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] Skipping network cache update for instance because it is Building. {{(pid=63531) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 651.216914] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] Skipping network cache update for instance because it is Building. {{(pid=63531) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 651.216914] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] Skipping network cache update for instance because it is Building. {{(pid=63531) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 651.216914] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Didn't find any instances for network info cache update. {{(pid=63531) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 651.218044] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 651.218247] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 651.218489] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 651.220992] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 651.221836] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 651.221836] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 651.221836] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63531) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 651.222511] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager.update_available_resource {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 651.276613] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Releasing lock "refresh_cache-56872ede-96f0-48ca-b289-86784f37527d" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 651.276852] env[63531]: DEBUG nova.compute.manager [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 651.277066] env[63531]: DEBUG nova.compute.manager [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 56872ede-96f0-48ca-b289-86784f37527d] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 651.277276] env[63531]: DEBUG nova.network.neutron [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 56872ede-96f0-48ca-b289-86784f37527d] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 651.311106] env[63531]: ERROR nova.compute.manager [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b82524d5-0c39-474f-b157-e64fef4fa205, please check neutron logs for more information. [ 651.311106] env[63531]: ERROR nova.compute.manager Traceback (most recent call last): [ 651.311106] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 651.311106] env[63531]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 651.311106] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 651.311106] env[63531]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 651.311106] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 651.311106] env[63531]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 651.311106] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 651.311106] env[63531]: ERROR nova.compute.manager self.force_reraise() [ 651.311106] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 651.311106] env[63531]: ERROR nova.compute.manager raise self.value [ 651.311106] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 651.311106] env[63531]: ERROR nova.compute.manager updated_port = self._update_port( [ 651.311106] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 651.311106] env[63531]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 651.311849] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 651.311849] env[63531]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 651.311849] env[63531]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b82524d5-0c39-474f-b157-e64fef4fa205, please check neutron logs for more information. [ 651.311849] env[63531]: ERROR nova.compute.manager [ 651.314756] env[63531]: Traceback (most recent call last): [ 651.314756] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 651.314756] env[63531]: listener.cb(fileno) [ 651.314756] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 651.314756] env[63531]: result = function(*args, **kwargs) [ 651.314756] env[63531]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 651.314756] env[63531]: return func(*args, **kwargs) [ 651.314756] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 651.314756] env[63531]: raise e [ 651.314756] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 651.314756] env[63531]: nwinfo = self.network_api.allocate_for_instance( [ 651.314756] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 651.314756] env[63531]: created_port_ids = self._update_ports_for_instance( [ 651.314756] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 651.314756] env[63531]: with excutils.save_and_reraise_exception(): [ 651.314756] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 651.314756] env[63531]: self.force_reraise() [ 651.314756] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 651.314756] env[63531]: raise self.value [ 651.314756] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 651.314756] env[63531]: updated_port = self._update_port( [ 651.314756] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 651.314756] env[63531]: _ensure_no_port_binding_failure(port) [ 651.314756] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 651.314756] env[63531]: raise exception.PortBindingFailed(port_id=port['id']) [ 651.314756] env[63531]: nova.exception.PortBindingFailed: Binding failed for port b82524d5-0c39-474f-b157-e64fef4fa205, please check neutron logs for more information. [ 651.314756] env[63531]: Removing descriptor: 20 [ 651.315450] env[63531]: ERROR nova.compute.manager [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b82524d5-0c39-474f-b157-e64fef4fa205, please check neutron logs for more information. [ 651.315450] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] Traceback (most recent call last): [ 651.315450] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 651.315450] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] yield resources [ 651.315450] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 651.315450] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] self.driver.spawn(context, instance, image_meta, [ 651.315450] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 651.315450] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 651.315450] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 651.315450] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] vm_ref = self.build_virtual_machine(instance, [ 651.315450] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 651.315784] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] vif_infos = vmwarevif.get_vif_info(self._session, [ 651.315784] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 651.315784] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] for vif in network_info: [ 651.315784] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 651.315784] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] return self._sync_wrapper(fn, *args, **kwargs) [ 651.315784] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 651.315784] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] self.wait() [ 651.315784] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 651.315784] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] self[:] = self._gt.wait() [ 651.315784] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 651.315784] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] return self._exit_event.wait() [ 651.315784] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 651.315784] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] result = hub.switch() [ 651.316100] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 651.316100] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] return self.greenlet.switch() [ 651.316100] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 651.316100] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] result = function(*args, **kwargs) [ 651.316100] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 651.316100] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] return func(*args, **kwargs) [ 651.316100] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 651.316100] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] raise e [ 651.316100] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 651.316100] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] nwinfo = self.network_api.allocate_for_instance( [ 651.316100] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 651.316100] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] created_port_ids = self._update_ports_for_instance( [ 651.316100] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 651.316541] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] with excutils.save_and_reraise_exception(): [ 651.316541] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 651.316541] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] self.force_reraise() [ 651.316541] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 651.316541] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] raise self.value [ 651.316541] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 651.316541] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] updated_port = self._update_port( [ 651.316541] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 651.316541] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] _ensure_no_port_binding_failure(port) [ 651.316541] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 651.316541] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] raise exception.PortBindingFailed(port_id=port['id']) [ 651.316541] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] nova.exception.PortBindingFailed: Binding failed for port b82524d5-0c39-474f-b157-e64fef4fa205, please check neutron logs for more information. [ 651.316541] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] [ 651.316856] env[63531]: INFO nova.compute.manager [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] Terminating instance [ 651.316856] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Acquiring lock "refresh_cache-316c12f7-bc05-4f70-98b4-579d4a7b69d0" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 651.316944] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Acquired lock "refresh_cache-316c12f7-bc05-4f70-98b4-579d4a7b69d0" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 651.317134] env[63531]: DEBUG nova.network.neutron [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 651.319644] env[63531]: DEBUG nova.network.neutron [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 56872ede-96f0-48ca-b289-86784f37527d] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 651.521755] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53a90627-8891-43a3-966d-e983e3dbf3f9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.529599] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3317f3e7-8c7a-40ed-a8f5-0c07642ffa6b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.564847] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b35c81f2-49f5-4d25-b768-0069c28ad44f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.574471] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efd28028-d414-410a-96f0-e5e5ed60aac6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.588445] env[63531]: DEBUG nova.compute.provider_tree [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 651.726133] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 651.823181] env[63531]: DEBUG nova.network.neutron [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 56872ede-96f0-48ca-b289-86784f37527d] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 651.839796] env[63531]: DEBUG nova.network.neutron [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 651.934384] env[63531]: DEBUG nova.network.neutron [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 651.973951] env[63531]: DEBUG nova.compute.manager [req-2428bb09-f8af-42f6-acba-7c4078919032 req-8f989a93-04e1-4fc0-a416-b62925aa4b79 service nova] [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] Received event network-changed-b82524d5-0c39-474f-b157-e64fef4fa205 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 651.974163] env[63531]: DEBUG nova.compute.manager [req-2428bb09-f8af-42f6-acba-7c4078919032 req-8f989a93-04e1-4fc0-a416-b62925aa4b79 service nova] [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] Refreshing instance network info cache due to event network-changed-b82524d5-0c39-474f-b157-e64fef4fa205. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 651.974381] env[63531]: DEBUG oslo_concurrency.lockutils [req-2428bb09-f8af-42f6-acba-7c4078919032 req-8f989a93-04e1-4fc0-a416-b62925aa4b79 service nova] Acquiring lock "refresh_cache-316c12f7-bc05-4f70-98b4-579d4a7b69d0" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 652.094413] env[63531]: DEBUG nova.scheduler.client.report [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 652.326495] env[63531]: INFO nova.compute.manager [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] [instance: 56872ede-96f0-48ca-b289-86784f37527d] Took 1.05 seconds to deallocate network for instance. [ 652.438910] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Releasing lock "refresh_cache-316c12f7-bc05-4f70-98b4-579d4a7b69d0" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 652.439412] env[63531]: DEBUG nova.compute.manager [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 652.439559] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 652.439894] env[63531]: DEBUG oslo_concurrency.lockutils [req-2428bb09-f8af-42f6-acba-7c4078919032 req-8f989a93-04e1-4fc0-a416-b62925aa4b79 service nova] Acquired lock "refresh_cache-316c12f7-bc05-4f70-98b4-579d4a7b69d0" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 652.440198] env[63531]: DEBUG nova.network.neutron [req-2428bb09-f8af-42f6-acba-7c4078919032 req-8f989a93-04e1-4fc0-a416-b62925aa4b79 service nova] [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] Refreshing network info cache for port b82524d5-0c39-474f-b157-e64fef4fa205 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 652.441154] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-17f24822-36fd-4846-8bde-1b074ca75220 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.456238] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-382c5661-22db-486b-b86b-e1e23b3a806d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.482209] env[63531]: WARNING nova.virt.vmwareapi.vmops [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 316c12f7-bc05-4f70-98b4-579d4a7b69d0 could not be found. [ 652.482209] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 652.482209] env[63531]: INFO nova.compute.manager [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] Took 0.04 seconds to destroy the instance on the hypervisor. [ 652.482819] env[63531]: DEBUG oslo.service.loopingcall [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 652.482819] env[63531]: DEBUG nova.compute.manager [-] [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 652.482819] env[63531]: DEBUG nova.network.neutron [-] [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 652.508036] env[63531]: DEBUG nova.network.neutron [-] [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 652.604036] env[63531]: DEBUG oslo_concurrency.lockutils [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.578s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 652.604036] env[63531]: DEBUG nova.compute.manager [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 652.605656] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.266s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 652.607283] env[63531]: INFO nova.compute.claims [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 652.965361] env[63531]: DEBUG nova.network.neutron [req-2428bb09-f8af-42f6-acba-7c4078919032 req-8f989a93-04e1-4fc0-a416-b62925aa4b79 service nova] [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 653.010838] env[63531]: DEBUG nova.network.neutron [-] [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 653.015833] env[63531]: DEBUG nova.network.neutron [req-2428bb09-f8af-42f6-acba-7c4078919032 req-8f989a93-04e1-4fc0-a416-b62925aa4b79 service nova] [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 653.114193] env[63531]: DEBUG nova.compute.utils [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 653.118164] env[63531]: DEBUG nova.compute.manager [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 653.118358] env[63531]: DEBUG nova.network.neutron [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 653.174864] env[63531]: DEBUG nova.policy [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a1734ce74e794662aa483a9d22264dd0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '123284e4b81d4f6e815a242aa9807566', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 653.356540] env[63531]: INFO nova.scheduler.client.report [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Deleted allocations for instance 56872ede-96f0-48ca-b289-86784f37527d [ 653.480284] env[63531]: DEBUG nova.network.neutron [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] Successfully created port: d408a90f-3083-499a-97c9-8090ac70d076 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 653.513335] env[63531]: INFO nova.compute.manager [-] [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] Took 1.03 seconds to deallocate network for instance. [ 653.517241] env[63531]: DEBUG nova.compute.claims [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] Aborting claim: {{(pid=63531) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 653.517241] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 653.518654] env[63531]: DEBUG oslo_concurrency.lockutils [req-2428bb09-f8af-42f6-acba-7c4078919032 req-8f989a93-04e1-4fc0-a416-b62925aa4b79 service nova] Releasing lock "refresh_cache-316c12f7-bc05-4f70-98b4-579d4a7b69d0" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 653.518897] env[63531]: DEBUG nova.compute.manager [req-2428bb09-f8af-42f6-acba-7c4078919032 req-8f989a93-04e1-4fc0-a416-b62925aa4b79 service nova] [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] Received event network-vif-deleted-b82524d5-0c39-474f-b157-e64fef4fa205 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 653.619516] env[63531]: DEBUG nova.compute.manager [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 653.807634] env[63531]: DEBUG oslo_concurrency.lockutils [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquiring lock "8dec0cac-defc-42f2-ab56-6b3ae60ad858" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 653.807860] env[63531]: DEBUG oslo_concurrency.lockutils [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "8dec0cac-defc-42f2-ab56-6b3ae60ad858" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 653.867983] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7b6271f5-ca4c-4524-b6a6-a1b9e0d44487 tempest-ListServerFiltersTestJSON-721720116 tempest-ListServerFiltersTestJSON-721720116-project-member] Lock "56872ede-96f0-48ca-b289-86784f37527d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 85.422s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 654.084933] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faf7715b-3b38-4b72-88c6-af6e0c477ba1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.093401] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3257444a-238d-44a4-9c6f-6a3ef937c379 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.125019] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2f29c26-6cbf-49e8-a556-72bae944dac5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.136992] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34b228bb-1939-418b-9d83-e12d9ae70ed9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.151631] env[63531]: DEBUG nova.compute.provider_tree [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 654.253062] env[63531]: DEBUG nova.compute.manager [req-376f6c94-4095-41f4-9d33-b1f993aa5067 req-16555f18-f307-4c23-92d9-be2981e8b56d service nova] [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] Received event network-changed-d408a90f-3083-499a-97c9-8090ac70d076 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 654.253062] env[63531]: DEBUG nova.compute.manager [req-376f6c94-4095-41f4-9d33-b1f993aa5067 req-16555f18-f307-4c23-92d9-be2981e8b56d service nova] [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] Refreshing instance network info cache due to event network-changed-d408a90f-3083-499a-97c9-8090ac70d076. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 654.253062] env[63531]: DEBUG oslo_concurrency.lockutils [req-376f6c94-4095-41f4-9d33-b1f993aa5067 req-16555f18-f307-4c23-92d9-be2981e8b56d service nova] Acquiring lock "refresh_cache-180526ef-0e57-4834-bc35-1e8e086e7323" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 654.253062] env[63531]: DEBUG oslo_concurrency.lockutils [req-376f6c94-4095-41f4-9d33-b1f993aa5067 req-16555f18-f307-4c23-92d9-be2981e8b56d service nova] Acquired lock "refresh_cache-180526ef-0e57-4834-bc35-1e8e086e7323" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 654.253062] env[63531]: DEBUG nova.network.neutron [req-376f6c94-4095-41f4-9d33-b1f993aa5067 req-16555f18-f307-4c23-92d9-be2981e8b56d service nova] [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] Refreshing network info cache for port d408a90f-3083-499a-97c9-8090ac70d076 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 654.376228] env[63531]: DEBUG nova.compute.manager [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 654.410872] env[63531]: ERROR nova.compute.manager [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d408a90f-3083-499a-97c9-8090ac70d076, please check neutron logs for more information. [ 654.410872] env[63531]: ERROR nova.compute.manager Traceback (most recent call last): [ 654.410872] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 654.410872] env[63531]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 654.410872] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 654.410872] env[63531]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 654.410872] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 654.410872] env[63531]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 654.410872] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 654.410872] env[63531]: ERROR nova.compute.manager self.force_reraise() [ 654.410872] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 654.410872] env[63531]: ERROR nova.compute.manager raise self.value [ 654.410872] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 654.410872] env[63531]: ERROR nova.compute.manager updated_port = self._update_port( [ 654.410872] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 654.410872] env[63531]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 654.411357] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 654.411357] env[63531]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 654.411357] env[63531]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d408a90f-3083-499a-97c9-8090ac70d076, please check neutron logs for more information. [ 654.411357] env[63531]: ERROR nova.compute.manager [ 654.411357] env[63531]: Traceback (most recent call last): [ 654.411357] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 654.411357] env[63531]: listener.cb(fileno) [ 654.411357] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 654.411357] env[63531]: result = function(*args, **kwargs) [ 654.411357] env[63531]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 654.411357] env[63531]: return func(*args, **kwargs) [ 654.411357] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 654.411357] env[63531]: raise e [ 654.411357] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 654.411357] env[63531]: nwinfo = self.network_api.allocate_for_instance( [ 654.411357] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 654.411357] env[63531]: created_port_ids = self._update_ports_for_instance( [ 654.411357] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 654.411357] env[63531]: with excutils.save_and_reraise_exception(): [ 654.411357] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 654.411357] env[63531]: self.force_reraise() [ 654.411357] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 654.411357] env[63531]: raise self.value [ 654.411357] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 654.411357] env[63531]: updated_port = self._update_port( [ 654.411357] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 654.411357] env[63531]: _ensure_no_port_binding_failure(port) [ 654.411357] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 654.411357] env[63531]: raise exception.PortBindingFailed(port_id=port['id']) [ 654.412090] env[63531]: nova.exception.PortBindingFailed: Binding failed for port d408a90f-3083-499a-97c9-8090ac70d076, please check neutron logs for more information. [ 654.412090] env[63531]: Removing descriptor: 17 [ 654.635328] env[63531]: DEBUG nova.compute.manager [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 654.654681] env[63531]: DEBUG nova.scheduler.client.report [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 654.659768] env[63531]: DEBUG nova.virt.hardware [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 654.659998] env[63531]: DEBUG nova.virt.hardware [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 654.660191] env[63531]: DEBUG nova.virt.hardware [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 654.660376] env[63531]: DEBUG nova.virt.hardware [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 654.660553] env[63531]: DEBUG nova.virt.hardware [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 654.660728] env[63531]: DEBUG nova.virt.hardware [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 654.660935] env[63531]: DEBUG nova.virt.hardware [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 654.661101] env[63531]: DEBUG nova.virt.hardware [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 654.661268] env[63531]: DEBUG nova.virt.hardware [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 654.661431] env[63531]: DEBUG nova.virt.hardware [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 654.661598] env[63531]: DEBUG nova.virt.hardware [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 654.662443] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-159db199-63f6-4ed9-90f0-60e475be73f2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.672904] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cbee829-6837-4ecf-a59b-2a19cdef7467 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.686937] env[63531]: ERROR nova.compute.manager [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d408a90f-3083-499a-97c9-8090ac70d076, please check neutron logs for more information. [ 654.686937] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] Traceback (most recent call last): [ 654.686937] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 654.686937] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] yield resources [ 654.686937] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 654.686937] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] self.driver.spawn(context, instance, image_meta, [ 654.686937] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 654.686937] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] self._vmops.spawn(context, instance, image_meta, injected_files, [ 654.686937] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 654.686937] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] vm_ref = self.build_virtual_machine(instance, [ 654.686937] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 654.687289] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] vif_infos = vmwarevif.get_vif_info(self._session, [ 654.687289] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 654.687289] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] for vif in network_info: [ 654.687289] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 654.687289] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] return self._sync_wrapper(fn, *args, **kwargs) [ 654.687289] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 654.687289] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] self.wait() [ 654.687289] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 654.687289] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] self[:] = self._gt.wait() [ 654.687289] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 654.687289] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] return self._exit_event.wait() [ 654.687289] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 654.687289] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] current.throw(*self._exc) [ 654.687664] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 654.687664] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] result = function(*args, **kwargs) [ 654.687664] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 654.687664] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] return func(*args, **kwargs) [ 654.687664] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 654.687664] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] raise e [ 654.687664] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 654.687664] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] nwinfo = self.network_api.allocate_for_instance( [ 654.687664] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 654.687664] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] created_port_ids = self._update_ports_for_instance( [ 654.687664] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 654.687664] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] with excutils.save_and_reraise_exception(): [ 654.687664] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 654.688144] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] self.force_reraise() [ 654.688144] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 654.688144] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] raise self.value [ 654.688144] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 654.688144] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] updated_port = self._update_port( [ 654.688144] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 654.688144] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] _ensure_no_port_binding_failure(port) [ 654.688144] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 654.688144] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] raise exception.PortBindingFailed(port_id=port['id']) [ 654.688144] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] nova.exception.PortBindingFailed: Binding failed for port d408a90f-3083-499a-97c9-8090ac70d076, please check neutron logs for more information. [ 654.688144] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] [ 654.688144] env[63531]: INFO nova.compute.manager [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] Terminating instance [ 654.689281] env[63531]: DEBUG oslo_concurrency.lockutils [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] Acquiring lock "refresh_cache-180526ef-0e57-4834-bc35-1e8e086e7323" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 654.774501] env[63531]: DEBUG nova.network.neutron [req-376f6c94-4095-41f4-9d33-b1f993aa5067 req-16555f18-f307-4c23-92d9-be2981e8b56d service nova] [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 654.902561] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 655.028503] env[63531]: DEBUG nova.network.neutron [req-376f6c94-4095-41f4-9d33-b1f993aa5067 req-16555f18-f307-4c23-92d9-be2981e8b56d service nova] [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 655.166657] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.561s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 655.167103] env[63531]: DEBUG nova.compute.manager [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 655.172440] env[63531]: DEBUG oslo_concurrency.lockutils [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.800s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 655.174031] env[63531]: INFO nova.compute.claims [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 655.531581] env[63531]: DEBUG oslo_concurrency.lockutils [req-376f6c94-4095-41f4-9d33-b1f993aa5067 req-16555f18-f307-4c23-92d9-be2981e8b56d service nova] Releasing lock "refresh_cache-180526ef-0e57-4834-bc35-1e8e086e7323" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 655.531702] env[63531]: DEBUG oslo_concurrency.lockutils [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] Acquired lock "refresh_cache-180526ef-0e57-4834-bc35-1e8e086e7323" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 655.531952] env[63531]: DEBUG nova.network.neutron [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 655.681118] env[63531]: DEBUG nova.compute.utils [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 655.683262] env[63531]: DEBUG nova.compute.manager [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 655.688024] env[63531]: DEBUG nova.network.neutron [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 655.739177] env[63531]: DEBUG nova.policy [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '30c2b29064b44c48ad7e76d445c65317', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '04a4c814245346a9b73253a25a822fef', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 656.029300] env[63531]: DEBUG nova.network.neutron [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] Successfully created port: 10033ac0-55f0-4735-a120-7e7d5218a6eb {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 656.058945] env[63531]: DEBUG nova.network.neutron [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 656.185176] env[63531]: DEBUG nova.network.neutron [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.186676] env[63531]: DEBUG nova.compute.manager [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 656.301557] env[63531]: DEBUG nova.compute.manager [req-9c1f96e5-4db4-4268-9141-b984c8bbb295 req-04bb439b-aa09-4326-81b9-b1750d3b6e85 service nova] [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] Received event network-vif-deleted-d408a90f-3083-499a-97c9-8090ac70d076 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 656.620856] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-719aeb24-026d-41d3-adee-4e79b65c1bbf {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.628877] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9074885a-1aae-4b38-8b49-a12ee9f8f9ef {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.659996] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75954e53-07fe-4df2-a145-090e7797ae2c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.667941] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d64829fd-77a5-455f-be35-e0c8c163059d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.681484] env[63531]: DEBUG nova.compute.provider_tree [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 656.697020] env[63531]: DEBUG oslo_concurrency.lockutils [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] Releasing lock "refresh_cache-180526ef-0e57-4834-bc35-1e8e086e7323" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 656.698104] env[63531]: DEBUG nova.compute.manager [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 656.698348] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 656.699120] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5f0c0ce8-81a0-4219-9493-89160fcec130 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.707773] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4287fc56-d78d-4c5c-9bfc-06df894c9cf4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.735689] env[63531]: WARNING nova.virt.vmwareapi.vmops [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 180526ef-0e57-4834-bc35-1e8e086e7323 could not be found. [ 656.735921] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 656.736568] env[63531]: INFO nova.compute.manager [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] Took 0.04 seconds to destroy the instance on the hypervisor. [ 656.736568] env[63531]: DEBUG oslo.service.loopingcall [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 656.736700] env[63531]: DEBUG nova.compute.manager [-] [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 656.736738] env[63531]: DEBUG nova.network.neutron [-] [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 656.766624] env[63531]: DEBUG nova.network.neutron [-] [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 657.047171] env[63531]: ERROR nova.compute.manager [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 10033ac0-55f0-4735-a120-7e7d5218a6eb, please check neutron logs for more information. [ 657.047171] env[63531]: ERROR nova.compute.manager Traceback (most recent call last): [ 657.047171] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 657.047171] env[63531]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 657.047171] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 657.047171] env[63531]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 657.047171] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 657.047171] env[63531]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 657.047171] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 657.047171] env[63531]: ERROR nova.compute.manager self.force_reraise() [ 657.047171] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 657.047171] env[63531]: ERROR nova.compute.manager raise self.value [ 657.047171] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 657.047171] env[63531]: ERROR nova.compute.manager updated_port = self._update_port( [ 657.047171] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 657.047171] env[63531]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 657.048804] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 657.048804] env[63531]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 657.048804] env[63531]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 10033ac0-55f0-4735-a120-7e7d5218a6eb, please check neutron logs for more information. [ 657.048804] env[63531]: ERROR nova.compute.manager [ 657.048804] env[63531]: Traceback (most recent call last): [ 657.048804] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 657.048804] env[63531]: listener.cb(fileno) [ 657.048804] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 657.048804] env[63531]: result = function(*args, **kwargs) [ 657.048804] env[63531]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 657.048804] env[63531]: return func(*args, **kwargs) [ 657.048804] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 657.048804] env[63531]: raise e [ 657.048804] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 657.048804] env[63531]: nwinfo = self.network_api.allocate_for_instance( [ 657.048804] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 657.048804] env[63531]: created_port_ids = self._update_ports_for_instance( [ 657.048804] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 657.048804] env[63531]: with excutils.save_and_reraise_exception(): [ 657.048804] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 657.048804] env[63531]: self.force_reraise() [ 657.048804] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 657.048804] env[63531]: raise self.value [ 657.048804] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 657.048804] env[63531]: updated_port = self._update_port( [ 657.048804] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 657.048804] env[63531]: _ensure_no_port_binding_failure(port) [ 657.048804] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 657.048804] env[63531]: raise exception.PortBindingFailed(port_id=port['id']) [ 657.050215] env[63531]: nova.exception.PortBindingFailed: Binding failed for port 10033ac0-55f0-4735-a120-7e7d5218a6eb, please check neutron logs for more information. [ 657.050215] env[63531]: Removing descriptor: 17 [ 657.071472] env[63531]: DEBUG oslo_concurrency.lockutils [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Acquiring lock "2998fb3d-6677-4fb3-952a-c74d8c6b7b9f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 657.071700] env[63531]: DEBUG oslo_concurrency.lockutils [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Lock "2998fb3d-6677-4fb3-952a-c74d8c6b7b9f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 657.184891] env[63531]: DEBUG nova.scheduler.client.report [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 657.200821] env[63531]: DEBUG nova.compute.manager [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 657.233220] env[63531]: DEBUG nova.virt.hardware [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 657.233485] env[63531]: DEBUG nova.virt.hardware [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 657.233655] env[63531]: DEBUG nova.virt.hardware [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 657.233883] env[63531]: DEBUG nova.virt.hardware [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 657.234157] env[63531]: DEBUG nova.virt.hardware [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 657.234356] env[63531]: DEBUG nova.virt.hardware [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 657.234580] env[63531]: DEBUG nova.virt.hardware [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 657.234768] env[63531]: DEBUG nova.virt.hardware [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 657.234903] env[63531]: DEBUG nova.virt.hardware [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 657.238427] env[63531]: DEBUG nova.virt.hardware [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 657.238427] env[63531]: DEBUG nova.virt.hardware [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 657.238427] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bc28286-5d38-47dc-9e2d-7220cb1a9d85 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.246360] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da1aef83-f304-450e-ac2e-b67e0163c9dc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.264297] env[63531]: ERROR nova.compute.manager [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 10033ac0-55f0-4735-a120-7e7d5218a6eb, please check neutron logs for more information. [ 657.264297] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] Traceback (most recent call last): [ 657.264297] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 657.264297] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] yield resources [ 657.264297] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 657.264297] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] self.driver.spawn(context, instance, image_meta, [ 657.264297] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 657.264297] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] self._vmops.spawn(context, instance, image_meta, injected_files, [ 657.264297] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 657.264297] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] vm_ref = self.build_virtual_machine(instance, [ 657.264297] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 657.265144] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] vif_infos = vmwarevif.get_vif_info(self._session, [ 657.265144] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 657.265144] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] for vif in network_info: [ 657.265144] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 657.265144] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] return self._sync_wrapper(fn, *args, **kwargs) [ 657.265144] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 657.265144] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] self.wait() [ 657.265144] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 657.265144] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] self[:] = self._gt.wait() [ 657.265144] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 657.265144] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] return self._exit_event.wait() [ 657.265144] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 657.265144] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] current.throw(*self._exc) [ 657.265612] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 657.265612] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] result = function(*args, **kwargs) [ 657.265612] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 657.265612] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] return func(*args, **kwargs) [ 657.265612] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 657.265612] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] raise e [ 657.265612] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 657.265612] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] nwinfo = self.network_api.allocate_for_instance( [ 657.265612] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 657.265612] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] created_port_ids = self._update_ports_for_instance( [ 657.265612] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 657.265612] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] with excutils.save_and_reraise_exception(): [ 657.265612] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 657.266093] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] self.force_reraise() [ 657.266093] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 657.266093] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] raise self.value [ 657.266093] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 657.266093] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] updated_port = self._update_port( [ 657.266093] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 657.266093] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] _ensure_no_port_binding_failure(port) [ 657.266093] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 657.266093] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] raise exception.PortBindingFailed(port_id=port['id']) [ 657.266093] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] nova.exception.PortBindingFailed: Binding failed for port 10033ac0-55f0-4735-a120-7e7d5218a6eb, please check neutron logs for more information. [ 657.266093] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] [ 657.266093] env[63531]: INFO nova.compute.manager [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] Terminating instance [ 657.267481] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Acquiring lock "refresh_cache-beb4b334-7d87-4c25-8efe-df92c43f0d26" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 657.267660] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Acquired lock "refresh_cache-beb4b334-7d87-4c25-8efe-df92c43f0d26" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 657.267824] env[63531]: DEBUG nova.network.neutron [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 657.271243] env[63531]: DEBUG nova.network.neutron [-] [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 657.323668] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Acquiring lock "64fb5831-0789-482b-bb98-67b29868c4c9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 657.324487] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Lock "64fb5831-0789-482b-bb98-67b29868c4c9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 657.689952] env[63531]: DEBUG oslo_concurrency.lockutils [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.520s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 657.691048] env[63531]: DEBUG nova.compute.manager [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 657.693122] env[63531]: DEBUG oslo_concurrency.lockutils [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.166s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 657.772741] env[63531]: INFO nova.compute.manager [-] [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] Took 1.04 seconds to deallocate network for instance. [ 657.774819] env[63531]: DEBUG nova.compute.claims [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] Aborting claim: {{(pid=63531) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 657.774986] env[63531]: DEBUG oslo_concurrency.lockutils [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 657.785910] env[63531]: DEBUG nova.network.neutron [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 657.869666] env[63531]: DEBUG nova.network.neutron [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 658.197314] env[63531]: DEBUG nova.compute.utils [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 658.201670] env[63531]: DEBUG nova.compute.manager [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 658.201844] env[63531]: DEBUG nova.network.neutron [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 658.240489] env[63531]: DEBUG nova.policy [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bfe034010b6247dcbc1dffb2e427629a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9c883e8ba760434483611afaa16c0132', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 658.355515] env[63531]: DEBUG nova.compute.manager [req-190f4c59-bd66-44eb-bdd8-ed4e154d4749 req-6a6a717c-6d72-4fb2-b44d-b8a28ae9d9b4 service nova] [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] Received event network-changed-10033ac0-55f0-4735-a120-7e7d5218a6eb {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 658.355765] env[63531]: DEBUG nova.compute.manager [req-190f4c59-bd66-44eb-bdd8-ed4e154d4749 req-6a6a717c-6d72-4fb2-b44d-b8a28ae9d9b4 service nova] [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] Refreshing instance network info cache due to event network-changed-10033ac0-55f0-4735-a120-7e7d5218a6eb. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 658.356011] env[63531]: DEBUG oslo_concurrency.lockutils [req-190f4c59-bd66-44eb-bdd8-ed4e154d4749 req-6a6a717c-6d72-4fb2-b44d-b8a28ae9d9b4 service nova] Acquiring lock "refresh_cache-beb4b334-7d87-4c25-8efe-df92c43f0d26" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 658.373573] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Releasing lock "refresh_cache-beb4b334-7d87-4c25-8efe-df92c43f0d26" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 658.374030] env[63531]: DEBUG nova.compute.manager [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 658.374654] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 658.374654] env[63531]: DEBUG oslo_concurrency.lockutils [req-190f4c59-bd66-44eb-bdd8-ed4e154d4749 req-6a6a717c-6d72-4fb2-b44d-b8a28ae9d9b4 service nova] Acquired lock "refresh_cache-beb4b334-7d87-4c25-8efe-df92c43f0d26" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 658.374751] env[63531]: DEBUG nova.network.neutron [req-190f4c59-bd66-44eb-bdd8-ed4e154d4749 req-6a6a717c-6d72-4fb2-b44d-b8a28ae9d9b4 service nova] [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] Refreshing network info cache for port 10033ac0-55f0-4735-a120-7e7d5218a6eb {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 658.375998] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fd535a71-2fbd-4b22-bd1f-f550f09164a4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.386084] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-582273de-eeb4-48fb-80a0-52a5a3974340 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.411419] env[63531]: WARNING nova.virt.vmwareapi.vmops [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance beb4b334-7d87-4c25-8efe-df92c43f0d26 could not be found. [ 658.411664] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 658.411850] env[63531]: INFO nova.compute.manager [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] Took 0.04 seconds to destroy the instance on the hypervisor. [ 658.412103] env[63531]: DEBUG oslo.service.loopingcall [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 658.414788] env[63531]: DEBUG nova.compute.manager [-] [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 658.414788] env[63531]: DEBUG nova.network.neutron [-] [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 658.434616] env[63531]: DEBUG nova.network.neutron [-] [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 658.544018] env[63531]: DEBUG nova.network.neutron [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] Successfully created port: 8b9ee4fb-3390-4145-892c-5750b12ef4e9 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 658.577821] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aae7f711-1630-4a15-97d7-84b7bebec1b6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.585807] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98b49c63-08ef-48d6-853f-ca19092b66e2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.619018] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc4875ba-3841-43a2-8091-1178d3dc37a0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.624550] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14f8c713-dcaf-4a16-b318-d1bd50e1259d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.638240] env[63531]: DEBUG nova.compute.provider_tree [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 658.702293] env[63531]: DEBUG nova.compute.manager [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 658.898812] env[63531]: DEBUG nova.network.neutron [req-190f4c59-bd66-44eb-bdd8-ed4e154d4749 req-6a6a717c-6d72-4fb2-b44d-b8a28ae9d9b4 service nova] [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 658.939551] env[63531]: DEBUG nova.network.neutron [-] [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 659.013733] env[63531]: DEBUG nova.network.neutron [req-190f4c59-bd66-44eb-bdd8-ed4e154d4749 req-6a6a717c-6d72-4fb2-b44d-b8a28ae9d9b4 service nova] [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 659.146031] env[63531]: DEBUG nova.scheduler.client.report [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 659.207256] env[63531]: INFO nova.virt.block_device [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] Booting with volume 54e7daec-bc20-49c7-a68d-4347bcbb73af at /dev/sda [ 659.254488] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8e797b25-d50e-4d73-9c62-e7607c2019c8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.263299] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a6b1425-12e8-43af-9054-7cdfe467d792 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.283970] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6b75a169-b8e7-48fa-8d21-6cc5c7fe611b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.291936] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13740aa7-d44d-49a2-949c-663ed33b4169 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.314053] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0d2f05c-4c51-4eb8-a471-322530883488 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.320721] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42ba55cf-8828-45c7-8217-af9638980fd4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.335534] env[63531]: DEBUG nova.virt.block_device [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] Updating existing volume attachment record: 47bf1911-3b0c-4c57-a818-fb373ad67c02 {{(pid=63531) _volume_attach /opt/stack/nova/nova/virt/block_device.py:679}} [ 659.424245] env[63531]: ERROR nova.compute.manager [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8b9ee4fb-3390-4145-892c-5750b12ef4e9, please check neutron logs for more information. [ 659.424245] env[63531]: ERROR nova.compute.manager Traceback (most recent call last): [ 659.424245] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 659.424245] env[63531]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 659.424245] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 659.424245] env[63531]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 659.424245] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 659.424245] env[63531]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 659.424245] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 659.424245] env[63531]: ERROR nova.compute.manager self.force_reraise() [ 659.424245] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 659.424245] env[63531]: ERROR nova.compute.manager raise self.value [ 659.424245] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 659.424245] env[63531]: ERROR nova.compute.manager updated_port = self._update_port( [ 659.424245] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 659.424245] env[63531]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 659.424793] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 659.424793] env[63531]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 659.424793] env[63531]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8b9ee4fb-3390-4145-892c-5750b12ef4e9, please check neutron logs for more information. [ 659.424793] env[63531]: ERROR nova.compute.manager [ 659.424793] env[63531]: Traceback (most recent call last): [ 659.424793] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 659.424793] env[63531]: listener.cb(fileno) [ 659.424793] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 659.424793] env[63531]: result = function(*args, **kwargs) [ 659.424793] env[63531]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 659.424793] env[63531]: return func(*args, **kwargs) [ 659.424793] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 659.424793] env[63531]: raise e [ 659.424793] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 659.424793] env[63531]: nwinfo = self.network_api.allocate_for_instance( [ 659.424793] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 659.424793] env[63531]: created_port_ids = self._update_ports_for_instance( [ 659.424793] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 659.424793] env[63531]: with excutils.save_and_reraise_exception(): [ 659.424793] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 659.424793] env[63531]: self.force_reraise() [ 659.424793] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 659.424793] env[63531]: raise self.value [ 659.424793] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 659.424793] env[63531]: updated_port = self._update_port( [ 659.424793] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 659.424793] env[63531]: _ensure_no_port_binding_failure(port) [ 659.424793] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 659.424793] env[63531]: raise exception.PortBindingFailed(port_id=port['id']) [ 659.425549] env[63531]: nova.exception.PortBindingFailed: Binding failed for port 8b9ee4fb-3390-4145-892c-5750b12ef4e9, please check neutron logs for more information. [ 659.425549] env[63531]: Removing descriptor: 17 [ 659.442543] env[63531]: INFO nova.compute.manager [-] [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] Took 1.03 seconds to deallocate network for instance. [ 659.445147] env[63531]: DEBUG nova.compute.claims [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] Aborting claim: {{(pid=63531) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 659.445333] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 659.516304] env[63531]: DEBUG oslo_concurrency.lockutils [req-190f4c59-bd66-44eb-bdd8-ed4e154d4749 req-6a6a717c-6d72-4fb2-b44d-b8a28ae9d9b4 service nova] Releasing lock "refresh_cache-beb4b334-7d87-4c25-8efe-df92c43f0d26" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 659.517050] env[63531]: DEBUG nova.compute.manager [req-190f4c59-bd66-44eb-bdd8-ed4e154d4749 req-6a6a717c-6d72-4fb2-b44d-b8a28ae9d9b4 service nova] [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] Received event network-vif-deleted-10033ac0-55f0-4735-a120-7e7d5218a6eb {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 659.653040] env[63531]: DEBUG oslo_concurrency.lockutils [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.960s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 659.653713] env[63531]: ERROR nova.compute.manager [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 06b2989d-8fee-41e1-bb1c-931c524779fd, please check neutron logs for more information. [ 659.653713] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] Traceback (most recent call last): [ 659.653713] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 659.653713] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] self.driver.spawn(context, instance, image_meta, [ 659.653713] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 659.653713] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] self._vmops.spawn(context, instance, image_meta, injected_files, [ 659.653713] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 659.653713] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] vm_ref = self.build_virtual_machine(instance, [ 659.653713] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 659.653713] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] vif_infos = vmwarevif.get_vif_info(self._session, [ 659.653713] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 659.654069] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] for vif in network_info: [ 659.654069] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 659.654069] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] return self._sync_wrapper(fn, *args, **kwargs) [ 659.654069] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 659.654069] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] self.wait() [ 659.654069] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 659.654069] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] self[:] = self._gt.wait() [ 659.654069] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 659.654069] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] return self._exit_event.wait() [ 659.654069] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 659.654069] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] current.throw(*self._exc) [ 659.654069] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 659.654069] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] result = function(*args, **kwargs) [ 659.654441] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 659.654441] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] return func(*args, **kwargs) [ 659.654441] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 659.654441] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] raise e [ 659.654441] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 659.654441] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] nwinfo = self.network_api.allocate_for_instance( [ 659.654441] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 659.654441] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] created_port_ids = self._update_ports_for_instance( [ 659.654441] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 659.654441] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] with excutils.save_and_reraise_exception(): [ 659.654441] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 659.654441] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] self.force_reraise() [ 659.654441] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 659.655325] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] raise self.value [ 659.655325] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 659.655325] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] updated_port = self._update_port( [ 659.655325] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 659.655325] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] _ensure_no_port_binding_failure(port) [ 659.655325] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 659.655325] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] raise exception.PortBindingFailed(port_id=port['id']) [ 659.655325] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] nova.exception.PortBindingFailed: Binding failed for port 06b2989d-8fee-41e1-bb1c-931c524779fd, please check neutron logs for more information. [ 659.655325] env[63531]: ERROR nova.compute.manager [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] [ 659.655325] env[63531]: DEBUG nova.compute.utils [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] Binding failed for port 06b2989d-8fee-41e1-bb1c-931c524779fd, please check neutron logs for more information. {{(pid=63531) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 659.655737] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.227s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 659.657249] env[63531]: INFO nova.compute.claims [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 659.659966] env[63531]: DEBUG nova.compute.manager [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] Build of instance 046304ef-b46f-473f-aa5b-5932b5078c54 was re-scheduled: Binding failed for port 06b2989d-8fee-41e1-bb1c-931c524779fd, please check neutron logs for more information. {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 659.660288] env[63531]: DEBUG nova.compute.manager [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] Unplugging VIFs for instance {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 659.660512] env[63531]: DEBUG oslo_concurrency.lockutils [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] Acquiring lock "refresh_cache-046304ef-b46f-473f-aa5b-5932b5078c54" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 659.660661] env[63531]: DEBUG oslo_concurrency.lockutils [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] Acquired lock "refresh_cache-046304ef-b46f-473f-aa5b-5932b5078c54" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 659.661151] env[63531]: DEBUG nova.network.neutron [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 660.184531] env[63531]: DEBUG nova.network.neutron [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 660.268185] env[63531]: DEBUG nova.network.neutron [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 660.410012] env[63531]: DEBUG nova.compute.manager [req-92bd4811-26fe-4083-af4a-942cb5ce2177 req-22c8ea4d-f444-4b78-ac7e-3f17b1674cf0 service nova] [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] Received event network-changed-8b9ee4fb-3390-4145-892c-5750b12ef4e9 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 660.410620] env[63531]: DEBUG nova.compute.manager [req-92bd4811-26fe-4083-af4a-942cb5ce2177 req-22c8ea4d-f444-4b78-ac7e-3f17b1674cf0 service nova] [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] Refreshing instance network info cache due to event network-changed-8b9ee4fb-3390-4145-892c-5750b12ef4e9. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 660.410620] env[63531]: DEBUG oslo_concurrency.lockutils [req-92bd4811-26fe-4083-af4a-942cb5ce2177 req-22c8ea4d-f444-4b78-ac7e-3f17b1674cf0 service nova] Acquiring lock "refresh_cache-c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 660.410620] env[63531]: DEBUG oslo_concurrency.lockutils [req-92bd4811-26fe-4083-af4a-942cb5ce2177 req-22c8ea4d-f444-4b78-ac7e-3f17b1674cf0 service nova] Acquired lock "refresh_cache-c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 660.410922] env[63531]: DEBUG nova.network.neutron [req-92bd4811-26fe-4083-af4a-942cb5ce2177 req-22c8ea4d-f444-4b78-ac7e-3f17b1674cf0 service nova] [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] Refreshing network info cache for port 8b9ee4fb-3390-4145-892c-5750b12ef4e9 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 660.770675] env[63531]: DEBUG oslo_concurrency.lockutils [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] Releasing lock "refresh_cache-046304ef-b46f-473f-aa5b-5932b5078c54" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 660.770927] env[63531]: DEBUG nova.compute.manager [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 660.771133] env[63531]: DEBUG nova.compute.manager [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 660.771310] env[63531]: DEBUG nova.network.neutron [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 660.789989] env[63531]: DEBUG nova.network.neutron [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 660.927470] env[63531]: DEBUG nova.network.neutron [req-92bd4811-26fe-4083-af4a-942cb5ce2177 req-22c8ea4d-f444-4b78-ac7e-3f17b1674cf0 service nova] [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 660.999816] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0c5074f-bb31-4c43-ba2f-1dc7c3fbdb71 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.004681] env[63531]: DEBUG nova.network.neutron [req-92bd4811-26fe-4083-af4a-942cb5ce2177 req-22c8ea4d-f444-4b78-ac7e-3f17b1674cf0 service nova] [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 661.011023] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc98949d-1715-4315-9cb1-0daa0bbd52d8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.040508] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fde3eae-f8b8-4bfe-8808-53079d360543 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.047634] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7fa089b-20ba-4544-91a2-7536d0c62db4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.060507] env[63531]: DEBUG nova.compute.provider_tree [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 661.293131] env[63531]: DEBUG nova.network.neutron [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 661.441745] env[63531]: DEBUG nova.compute.manager [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 661.442273] env[63531]: DEBUG nova.virt.hardware [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 661.442486] env[63531]: DEBUG nova.virt.hardware [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 661.442640] env[63531]: DEBUG nova.virt.hardware [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 661.442820] env[63531]: DEBUG nova.virt.hardware [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 661.442963] env[63531]: DEBUG nova.virt.hardware [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 661.443128] env[63531]: DEBUG nova.virt.hardware [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 661.443330] env[63531]: DEBUG nova.virt.hardware [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 661.443490] env[63531]: DEBUG nova.virt.hardware [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 661.443650] env[63531]: DEBUG nova.virt.hardware [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 661.443809] env[63531]: DEBUG nova.virt.hardware [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 661.443978] env[63531]: DEBUG nova.virt.hardware [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 661.445151] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6228688-93af-41f7-847f-e4f638a20115 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.453584] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97149a5e-52c3-4b06-b507-278bafa8dcc5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.468103] env[63531]: ERROR nova.compute.manager [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8b9ee4fb-3390-4145-892c-5750b12ef4e9, please check neutron logs for more information. [ 661.468103] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] Traceback (most recent call last): [ 661.468103] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 661.468103] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] yield resources [ 661.468103] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 661.468103] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] self.driver.spawn(context, instance, image_meta, [ 661.468103] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 661.468103] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 661.468103] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 661.468103] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] vm_ref = self.build_virtual_machine(instance, [ 661.468103] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 661.468451] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] vif_infos = vmwarevif.get_vif_info(self._session, [ 661.468451] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 661.468451] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] for vif in network_info: [ 661.468451] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 661.468451] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] return self._sync_wrapper(fn, *args, **kwargs) [ 661.468451] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 661.468451] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] self.wait() [ 661.468451] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 661.468451] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] self[:] = self._gt.wait() [ 661.468451] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 661.468451] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] return self._exit_event.wait() [ 661.468451] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 661.468451] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] current.throw(*self._exc) [ 661.468812] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 661.468812] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] result = function(*args, **kwargs) [ 661.468812] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 661.468812] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] return func(*args, **kwargs) [ 661.468812] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 661.468812] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] raise e [ 661.468812] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 661.468812] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] nwinfo = self.network_api.allocate_for_instance( [ 661.468812] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 661.468812] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] created_port_ids = self._update_ports_for_instance( [ 661.468812] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 661.468812] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] with excutils.save_and_reraise_exception(): [ 661.468812] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 661.469176] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] self.force_reraise() [ 661.469176] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 661.469176] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] raise self.value [ 661.469176] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 661.469176] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] updated_port = self._update_port( [ 661.469176] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 661.469176] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] _ensure_no_port_binding_failure(port) [ 661.469176] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 661.469176] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] raise exception.PortBindingFailed(port_id=port['id']) [ 661.469176] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] nova.exception.PortBindingFailed: Binding failed for port 8b9ee4fb-3390-4145-892c-5750b12ef4e9, please check neutron logs for more information. [ 661.469176] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] [ 661.469176] env[63531]: INFO nova.compute.manager [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] Terminating instance [ 661.470390] env[63531]: DEBUG oslo_concurrency.lockutils [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] Acquiring lock "refresh_cache-c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 661.507055] env[63531]: DEBUG oslo_concurrency.lockutils [req-92bd4811-26fe-4083-af4a-942cb5ce2177 req-22c8ea4d-f444-4b78-ac7e-3f17b1674cf0 service nova] Releasing lock "refresh_cache-c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 661.507315] env[63531]: DEBUG nova.compute.manager [req-92bd4811-26fe-4083-af4a-942cb5ce2177 req-22c8ea4d-f444-4b78-ac7e-3f17b1674cf0 service nova] [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] Received event network-vif-deleted-8b9ee4fb-3390-4145-892c-5750b12ef4e9 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 661.507674] env[63531]: DEBUG oslo_concurrency.lockutils [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] Acquired lock "refresh_cache-c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 661.507838] env[63531]: DEBUG nova.network.neutron [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 661.563811] env[63531]: DEBUG nova.scheduler.client.report [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 661.796031] env[63531]: INFO nova.compute.manager [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] [instance: 046304ef-b46f-473f-aa5b-5932b5078c54] Took 1.02 seconds to deallocate network for instance. [ 662.029808] env[63531]: DEBUG nova.network.neutron [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 662.069657] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.414s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 662.070168] env[63531]: DEBUG nova.compute.manager [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 662.072892] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.539s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 662.075045] env[63531]: INFO nova.compute.claims [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 662.115854] env[63531]: DEBUG nova.network.neutron [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 662.578656] env[63531]: DEBUG nova.compute.utils [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 662.583059] env[63531]: DEBUG nova.compute.manager [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 662.583059] env[63531]: DEBUG nova.network.neutron [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 662.618563] env[63531]: DEBUG oslo_concurrency.lockutils [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] Releasing lock "refresh_cache-c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 662.619113] env[63531]: DEBUG nova.compute.manager [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 662.619433] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8a2cd6c0-ce5a-45a0-a1ff-4145af456cfa {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.622697] env[63531]: DEBUG nova.policy [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e9d80dd581b8420b8ebf4c74a4728220', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a9a802cef77b49c1945fb1a3c05ee3a8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 662.630381] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-129bf37a-9c66-4da1-aadd-a6621a31c406 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.653112] env[63531]: WARNING nova.virt.vmwareapi.driver [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f could not be found. [ 662.653112] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 662.653353] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-68cb54f7-25be-4406-94dd-febfa69c6d30 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.661302] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-717eb02c-f375-4492-a5e7-58a6f539e3cd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.681839] env[63531]: WARNING nova.virt.vmwareapi.vmops [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f could not be found. [ 662.682079] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 662.682271] env[63531]: INFO nova.compute.manager [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] Took 0.06 seconds to destroy the instance on the hypervisor. [ 662.682522] env[63531]: DEBUG oslo.service.loopingcall [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 662.682741] env[63531]: DEBUG nova.compute.manager [-] [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 662.682833] env[63531]: DEBUG nova.network.neutron [-] [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 662.701032] env[63531]: DEBUG nova.network.neutron [-] [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 662.832539] env[63531]: INFO nova.scheduler.client.report [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] Deleted allocations for instance 046304ef-b46f-473f-aa5b-5932b5078c54 [ 662.915526] env[63531]: DEBUG nova.network.neutron [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] Successfully created port: 0f7134ca-5e5e-4a9d-9380-d19042daa7b8 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 663.083147] env[63531]: DEBUG nova.compute.manager [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 663.203664] env[63531]: DEBUG nova.network.neutron [-] [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 663.344548] env[63531]: DEBUG oslo_concurrency.lockutils [None req-d8859582-e392-4743-9245-63fe495e967b tempest-FloatingIPsAssociationTestJSON-527631555 tempest-FloatingIPsAssociationTestJSON-527631555-project-member] Lock "046304ef-b46f-473f-aa5b-5932b5078c54" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 86.877s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 663.494819] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6974592b-c6f7-44c2-84e8-599380564919 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.503762] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-254a0f24-a534-4861-82bd-37b2a693b139 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.535034] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d0b5f8e-8f40-498b-a367-92762111a420 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.543867] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a01e42c-c521-40c2-8e00-458b1eaa6d05 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.559192] env[63531]: DEBUG nova.compute.provider_tree [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 663.587193] env[63531]: DEBUG nova.compute.manager [req-1dcbe710-3f6c-4532-affe-5659bb26ca6f req-dcbf35bc-cc90-4deb-ad8a-743a1d815e4a service nova] [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] Received event network-changed-0f7134ca-5e5e-4a9d-9380-d19042daa7b8 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 663.587639] env[63531]: DEBUG nova.compute.manager [req-1dcbe710-3f6c-4532-affe-5659bb26ca6f req-dcbf35bc-cc90-4deb-ad8a-743a1d815e4a service nova] [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] Refreshing instance network info cache due to event network-changed-0f7134ca-5e5e-4a9d-9380-d19042daa7b8. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 663.587639] env[63531]: DEBUG oslo_concurrency.lockutils [req-1dcbe710-3f6c-4532-affe-5659bb26ca6f req-dcbf35bc-cc90-4deb-ad8a-743a1d815e4a service nova] Acquiring lock "refresh_cache-5521c3ab-d017-473d-b8ce-d37c8b2bcc8f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 663.587766] env[63531]: DEBUG oslo_concurrency.lockutils [req-1dcbe710-3f6c-4532-affe-5659bb26ca6f req-dcbf35bc-cc90-4deb-ad8a-743a1d815e4a service nova] Acquired lock "refresh_cache-5521c3ab-d017-473d-b8ce-d37c8b2bcc8f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 663.587923] env[63531]: DEBUG nova.network.neutron [req-1dcbe710-3f6c-4532-affe-5659bb26ca6f req-dcbf35bc-cc90-4deb-ad8a-743a1d815e4a service nova] [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] Refreshing network info cache for port 0f7134ca-5e5e-4a9d-9380-d19042daa7b8 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 663.706944] env[63531]: INFO nova.compute.manager [-] [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] Took 1.02 seconds to deallocate network for instance. [ 663.769907] env[63531]: ERROR nova.compute.manager [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0f7134ca-5e5e-4a9d-9380-d19042daa7b8, please check neutron logs for more information. [ 663.769907] env[63531]: ERROR nova.compute.manager Traceback (most recent call last): [ 663.769907] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 663.769907] env[63531]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 663.769907] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 663.769907] env[63531]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 663.769907] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 663.769907] env[63531]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 663.769907] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 663.769907] env[63531]: ERROR nova.compute.manager self.force_reraise() [ 663.769907] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 663.769907] env[63531]: ERROR nova.compute.manager raise self.value [ 663.769907] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 663.769907] env[63531]: ERROR nova.compute.manager updated_port = self._update_port( [ 663.769907] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 663.769907] env[63531]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 663.770352] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 663.770352] env[63531]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 663.770352] env[63531]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0f7134ca-5e5e-4a9d-9380-d19042daa7b8, please check neutron logs for more information. [ 663.770352] env[63531]: ERROR nova.compute.manager [ 663.770352] env[63531]: Traceback (most recent call last): [ 663.770352] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 663.770352] env[63531]: listener.cb(fileno) [ 663.770352] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 663.770352] env[63531]: result = function(*args, **kwargs) [ 663.770352] env[63531]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 663.770352] env[63531]: return func(*args, **kwargs) [ 663.770352] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 663.770352] env[63531]: raise e [ 663.770352] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 663.770352] env[63531]: nwinfo = self.network_api.allocate_for_instance( [ 663.770352] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 663.770352] env[63531]: created_port_ids = self._update_ports_for_instance( [ 663.770352] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 663.770352] env[63531]: with excutils.save_and_reraise_exception(): [ 663.770352] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 663.770352] env[63531]: self.force_reraise() [ 663.770352] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 663.770352] env[63531]: raise self.value [ 663.770352] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 663.770352] env[63531]: updated_port = self._update_port( [ 663.770352] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 663.770352] env[63531]: _ensure_no_port_binding_failure(port) [ 663.770352] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 663.770352] env[63531]: raise exception.PortBindingFailed(port_id=port['id']) [ 663.771014] env[63531]: nova.exception.PortBindingFailed: Binding failed for port 0f7134ca-5e5e-4a9d-9380-d19042daa7b8, please check neutron logs for more information. [ 663.771014] env[63531]: Removing descriptor: 17 [ 663.849210] env[63531]: DEBUG nova.compute.manager [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: 3f3113dd-863f-442d-a214-28dde377d227] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 664.062400] env[63531]: DEBUG nova.scheduler.client.report [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 664.098952] env[63531]: DEBUG nova.compute.manager [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 664.120090] env[63531]: DEBUG nova.network.neutron [req-1dcbe710-3f6c-4532-affe-5659bb26ca6f req-dcbf35bc-cc90-4deb-ad8a-743a1d815e4a service nova] [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 664.129579] env[63531]: DEBUG nova.virt.hardware [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 664.131740] env[63531]: DEBUG nova.virt.hardware [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 664.131740] env[63531]: DEBUG nova.virt.hardware [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 664.131740] env[63531]: DEBUG nova.virt.hardware [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 664.131740] env[63531]: DEBUG nova.virt.hardware [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 664.131740] env[63531]: DEBUG nova.virt.hardware [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 664.131961] env[63531]: DEBUG nova.virt.hardware [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 664.131961] env[63531]: DEBUG nova.virt.hardware [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 664.131961] env[63531]: DEBUG nova.virt.hardware [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 664.131961] env[63531]: DEBUG nova.virt.hardware [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 664.131961] env[63531]: DEBUG nova.virt.hardware [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 664.132245] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48bac36e-c877-4130-8008-870ce3831c61 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.140293] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a67c04c6-5bd4-4dd5-b6af-530b96f309e3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.157985] env[63531]: ERROR nova.compute.manager [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0f7134ca-5e5e-4a9d-9380-d19042daa7b8, please check neutron logs for more information. [ 664.157985] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] Traceback (most recent call last): [ 664.157985] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 664.157985] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] yield resources [ 664.157985] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 664.157985] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] self.driver.spawn(context, instance, image_meta, [ 664.157985] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 664.157985] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 664.157985] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 664.157985] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] vm_ref = self.build_virtual_machine(instance, [ 664.157985] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 664.158370] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] vif_infos = vmwarevif.get_vif_info(self._session, [ 664.158370] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 664.158370] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] for vif in network_info: [ 664.158370] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 664.158370] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] return self._sync_wrapper(fn, *args, **kwargs) [ 664.158370] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 664.158370] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] self.wait() [ 664.158370] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 664.158370] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] self[:] = self._gt.wait() [ 664.158370] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 664.158370] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] return self._exit_event.wait() [ 664.158370] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 664.158370] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] current.throw(*self._exc) [ 664.158722] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 664.158722] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] result = function(*args, **kwargs) [ 664.158722] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 664.158722] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] return func(*args, **kwargs) [ 664.158722] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 664.158722] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] raise e [ 664.158722] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 664.158722] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] nwinfo = self.network_api.allocate_for_instance( [ 664.158722] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 664.158722] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] created_port_ids = self._update_ports_for_instance( [ 664.158722] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 664.158722] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] with excutils.save_and_reraise_exception(): [ 664.158722] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 664.159090] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] self.force_reraise() [ 664.159090] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 664.159090] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] raise self.value [ 664.159090] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 664.159090] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] updated_port = self._update_port( [ 664.159090] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 664.159090] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] _ensure_no_port_binding_failure(port) [ 664.159090] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 664.159090] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] raise exception.PortBindingFailed(port_id=port['id']) [ 664.159090] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] nova.exception.PortBindingFailed: Binding failed for port 0f7134ca-5e5e-4a9d-9380-d19042daa7b8, please check neutron logs for more information. [ 664.159090] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] [ 664.159090] env[63531]: INFO nova.compute.manager [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] Terminating instance [ 664.163125] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] Acquiring lock "refresh_cache-5521c3ab-d017-473d-b8ce-d37c8b2bcc8f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 664.201581] env[63531]: DEBUG nova.network.neutron [req-1dcbe710-3f6c-4532-affe-5659bb26ca6f req-dcbf35bc-cc90-4deb-ad8a-743a1d815e4a service nova] [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 664.273296] env[63531]: INFO nova.compute.manager [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] Took 0.57 seconds to detach 1 volumes for instance. [ 664.276668] env[63531]: DEBUG nova.compute.claims [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] Aborting claim: {{(pid=63531) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 664.276668] env[63531]: DEBUG oslo_concurrency.lockutils [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 664.373432] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 664.570732] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.498s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 664.571345] env[63531]: DEBUG nova.compute.manager [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 664.574527] env[63531]: DEBUG oslo_concurrency.lockutils [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.915s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 664.577275] env[63531]: INFO nova.compute.claims [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 664.707661] env[63531]: DEBUG oslo_concurrency.lockutils [req-1dcbe710-3f6c-4532-affe-5659bb26ca6f req-dcbf35bc-cc90-4deb-ad8a-743a1d815e4a service nova] Releasing lock "refresh_cache-5521c3ab-d017-473d-b8ce-d37c8b2bcc8f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 664.707661] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] Acquired lock "refresh_cache-5521c3ab-d017-473d-b8ce-d37c8b2bcc8f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 664.707661] env[63531]: DEBUG nova.network.neutron [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 665.087034] env[63531]: DEBUG nova.compute.utils [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 665.088520] env[63531]: DEBUG nova.compute.manager [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 665.088520] env[63531]: DEBUG nova.network.neutron [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 665.158397] env[63531]: DEBUG nova.policy [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9cf1752c51cd476d8a2c6bb04e2e16b5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '729a0934b225484f9de897f8714875c7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 665.232747] env[63531]: DEBUG nova.network.neutron [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 665.361761] env[63531]: DEBUG nova.network.neutron [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 665.503217] env[63531]: DEBUG nova.network.neutron [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] Successfully created port: 1baeb6e8-6746-4bad-b9da-bce60d8935d5 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 665.591769] env[63531]: DEBUG nova.compute.manager [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 665.617646] env[63531]: DEBUG nova.compute.manager [req-47501d47-dc5c-4ea9-a408-0a4e0b7b2d2d req-c1f3f269-a477-438c-a672-a97279b2a277 service nova] [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] Received event network-vif-deleted-0f7134ca-5e5e-4a9d-9380-d19042daa7b8 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 665.864068] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] Releasing lock "refresh_cache-5521c3ab-d017-473d-b8ce-d37c8b2bcc8f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 665.864577] env[63531]: DEBUG nova.compute.manager [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 665.864781] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 665.865089] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1192d0d7-7ffe-4004-b70a-e656bbaf8bdc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.877414] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-996a2ec0-4887-4e54-a213-0a480f1ac181 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.906175] env[63531]: WARNING nova.virt.vmwareapi.vmops [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f could not be found. [ 665.906558] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 665.906809] env[63531]: INFO nova.compute.manager [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] Took 0.04 seconds to destroy the instance on the hypervisor. [ 665.907130] env[63531]: DEBUG oslo.service.loopingcall [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 665.913649] env[63531]: DEBUG nova.compute.manager [-] [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 665.913882] env[63531]: DEBUG nova.network.neutron [-] [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 665.930854] env[63531]: DEBUG nova.network.neutron [-] [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 666.032499] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b55775c8-8bc0-4015-ae5f-931919f681cf {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.040031] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a78462b-79b5-4394-aeda-678195585dba {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.069779] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1613421d-f66a-4d0e-a7c5-6fc8b5d74d27 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.077112] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b0328b7-9fc0-4723-8420-4e91b601ffe8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.091016] env[63531]: DEBUG nova.compute.provider_tree [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 666.413519] env[63531]: ERROR nova.compute.manager [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1baeb6e8-6746-4bad-b9da-bce60d8935d5, please check neutron logs for more information. [ 666.413519] env[63531]: ERROR nova.compute.manager Traceback (most recent call last): [ 666.413519] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 666.413519] env[63531]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 666.413519] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 666.413519] env[63531]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 666.413519] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 666.413519] env[63531]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 666.413519] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 666.413519] env[63531]: ERROR nova.compute.manager self.force_reraise() [ 666.413519] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 666.413519] env[63531]: ERROR nova.compute.manager raise self.value [ 666.413519] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 666.413519] env[63531]: ERROR nova.compute.manager updated_port = self._update_port( [ 666.413519] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 666.413519] env[63531]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 666.414044] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 666.414044] env[63531]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 666.414044] env[63531]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1baeb6e8-6746-4bad-b9da-bce60d8935d5, please check neutron logs for more information. [ 666.414044] env[63531]: ERROR nova.compute.manager [ 666.417382] env[63531]: Traceback (most recent call last): [ 666.417382] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 666.417382] env[63531]: listener.cb(fileno) [ 666.417382] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 666.417382] env[63531]: result = function(*args, **kwargs) [ 666.417382] env[63531]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 666.417382] env[63531]: return func(*args, **kwargs) [ 666.417382] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 666.417382] env[63531]: raise e [ 666.417382] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 666.417382] env[63531]: nwinfo = self.network_api.allocate_for_instance( [ 666.417382] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 666.417382] env[63531]: created_port_ids = self._update_ports_for_instance( [ 666.417382] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 666.417382] env[63531]: with excutils.save_and_reraise_exception(): [ 666.417382] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 666.417382] env[63531]: self.force_reraise() [ 666.417382] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 666.417382] env[63531]: raise self.value [ 666.417382] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 666.417382] env[63531]: updated_port = self._update_port( [ 666.417382] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 666.417382] env[63531]: _ensure_no_port_binding_failure(port) [ 666.417382] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 666.417382] env[63531]: raise exception.PortBindingFailed(port_id=port['id']) [ 666.417382] env[63531]: nova.exception.PortBindingFailed: Binding failed for port 1baeb6e8-6746-4bad-b9da-bce60d8935d5, please check neutron logs for more information. [ 666.417382] env[63531]: Removing descriptor: 17 [ 666.436844] env[63531]: DEBUG nova.network.neutron [-] [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 666.594093] env[63531]: DEBUG nova.scheduler.client.report [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 666.601666] env[63531]: DEBUG nova.compute.manager [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 666.629049] env[63531]: DEBUG nova.virt.hardware [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 666.629183] env[63531]: DEBUG nova.virt.hardware [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 666.629358] env[63531]: DEBUG nova.virt.hardware [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 666.629563] env[63531]: DEBUG nova.virt.hardware [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 666.629756] env[63531]: DEBUG nova.virt.hardware [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 666.629852] env[63531]: DEBUG nova.virt.hardware [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 666.630076] env[63531]: DEBUG nova.virt.hardware [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 666.630238] env[63531]: DEBUG nova.virt.hardware [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 666.630402] env[63531]: DEBUG nova.virt.hardware [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 666.630566] env[63531]: DEBUG nova.virt.hardware [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 666.630737] env[63531]: DEBUG nova.virt.hardware [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 666.631862] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7ea2754-0bc3-4575-8ded-8bf45cedef1f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.640768] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-760a4829-ebfc-4f13-815c-297e8d4d73b2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.655492] env[63531]: ERROR nova.compute.manager [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1baeb6e8-6746-4bad-b9da-bce60d8935d5, please check neutron logs for more information. [ 666.655492] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] Traceback (most recent call last): [ 666.655492] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 666.655492] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] yield resources [ 666.655492] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 666.655492] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] self.driver.spawn(context, instance, image_meta, [ 666.655492] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 666.655492] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 666.655492] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 666.655492] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] vm_ref = self.build_virtual_machine(instance, [ 666.655492] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 666.655902] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] vif_infos = vmwarevif.get_vif_info(self._session, [ 666.655902] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 666.655902] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] for vif in network_info: [ 666.655902] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 666.655902] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] return self._sync_wrapper(fn, *args, **kwargs) [ 666.655902] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 666.655902] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] self.wait() [ 666.655902] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 666.655902] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] self[:] = self._gt.wait() [ 666.655902] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 666.655902] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] return self._exit_event.wait() [ 666.655902] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 666.655902] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] current.throw(*self._exc) [ 666.656270] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 666.656270] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] result = function(*args, **kwargs) [ 666.656270] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 666.656270] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] return func(*args, **kwargs) [ 666.656270] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 666.656270] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] raise e [ 666.656270] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 666.656270] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] nwinfo = self.network_api.allocate_for_instance( [ 666.656270] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 666.656270] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] created_port_ids = self._update_ports_for_instance( [ 666.656270] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 666.656270] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] with excutils.save_and_reraise_exception(): [ 666.656270] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 666.656662] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] self.force_reraise() [ 666.656662] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 666.656662] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] raise self.value [ 666.656662] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 666.656662] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] updated_port = self._update_port( [ 666.656662] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 666.656662] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] _ensure_no_port_binding_failure(port) [ 666.656662] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 666.656662] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] raise exception.PortBindingFailed(port_id=port['id']) [ 666.656662] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] nova.exception.PortBindingFailed: Binding failed for port 1baeb6e8-6746-4bad-b9da-bce60d8935d5, please check neutron logs for more information. [ 666.656662] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] [ 666.656662] env[63531]: INFO nova.compute.manager [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] Terminating instance [ 666.657931] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] Acquiring lock "refresh_cache-9a4f813b-5aac-4458-8fc3-af33b4b9636c" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 666.658114] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] Acquired lock "refresh_cache-9a4f813b-5aac-4458-8fc3-af33b4b9636c" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 666.658285] env[63531]: DEBUG nova.network.neutron [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 666.939568] env[63531]: INFO nova.compute.manager [-] [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] Took 1.03 seconds to deallocate network for instance. [ 666.942269] env[63531]: DEBUG nova.compute.claims [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] Aborting claim: {{(pid=63531) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 666.942450] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 667.099718] env[63531]: DEBUG oslo_concurrency.lockutils [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.525s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 667.100268] env[63531]: DEBUG nova.compute.manager [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 667.104237] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.917s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 667.176771] env[63531]: DEBUG nova.network.neutron [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 667.267889] env[63531]: DEBUG nova.network.neutron [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 667.611824] env[63531]: DEBUG nova.compute.utils [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 667.613200] env[63531]: DEBUG nova.compute.manager [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 667.613410] env[63531]: DEBUG nova.network.neutron [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 667.643917] env[63531]: DEBUG nova.compute.manager [req-f08a1e88-8879-41e0-9e81-ad5f82f77478 req-911f706d-70f4-42b4-ab05-fe1e5c4a2890 service nova] [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] Received event network-changed-1baeb6e8-6746-4bad-b9da-bce60d8935d5 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 667.644139] env[63531]: DEBUG nova.compute.manager [req-f08a1e88-8879-41e0-9e81-ad5f82f77478 req-911f706d-70f4-42b4-ab05-fe1e5c4a2890 service nova] [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] Refreshing instance network info cache due to event network-changed-1baeb6e8-6746-4bad-b9da-bce60d8935d5. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 667.644348] env[63531]: DEBUG oslo_concurrency.lockutils [req-f08a1e88-8879-41e0-9e81-ad5f82f77478 req-911f706d-70f4-42b4-ab05-fe1e5c4a2890 service nova] Acquiring lock "refresh_cache-9a4f813b-5aac-4458-8fc3-af33b4b9636c" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 667.656881] env[63531]: DEBUG nova.policy [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '268494bb8db84ede854304a23497d130', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a0e384d6cf184a98aa3602eca6b3f410', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 667.773542] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] Releasing lock "refresh_cache-9a4f813b-5aac-4458-8fc3-af33b4b9636c" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 667.773972] env[63531]: DEBUG nova.compute.manager [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 667.774220] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 667.774552] env[63531]: DEBUG oslo_concurrency.lockutils [req-f08a1e88-8879-41e0-9e81-ad5f82f77478 req-911f706d-70f4-42b4-ab05-fe1e5c4a2890 service nova] Acquired lock "refresh_cache-9a4f813b-5aac-4458-8fc3-af33b4b9636c" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 667.774728] env[63531]: DEBUG nova.network.neutron [req-f08a1e88-8879-41e0-9e81-ad5f82f77478 req-911f706d-70f4-42b4-ab05-fe1e5c4a2890 service nova] [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] Refreshing network info cache for port 1baeb6e8-6746-4bad-b9da-bce60d8935d5 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 667.775783] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c9f404e7-e989-45df-95fa-d53222f1edd2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.786277] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b724ffb-cc87-4f29-bc73-0e1567511d91 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.814850] env[63531]: WARNING nova.virt.vmwareapi.vmops [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9a4f813b-5aac-4458-8fc3-af33b4b9636c could not be found. [ 667.815091] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 667.815276] env[63531]: INFO nova.compute.manager [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] Took 0.04 seconds to destroy the instance on the hypervisor. [ 667.815510] env[63531]: DEBUG oslo.service.loopingcall [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 667.817994] env[63531]: DEBUG nova.compute.manager [-] [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 667.818100] env[63531]: DEBUG nova.network.neutron [-] [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 667.833883] env[63531]: DEBUG nova.network.neutron [-] [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 667.954821] env[63531]: DEBUG nova.network.neutron [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] Successfully created port: 50ecfd39-5844-4ad6-b4c9-f29416ae5693 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 667.991291] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f63a4ae4-a366-4dac-83e5-a61d17d45a8e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.998985] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-600b851f-1ff2-4baa-ba0f-9635a28c61c0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.031106] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e0918a3-ced0-4197-9c64-58b216a80e4d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.040376] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc8030c4-3806-426f-93ec-6e09967644bb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.055977] env[63531]: DEBUG nova.compute.provider_tree [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 668.121152] env[63531]: DEBUG nova.compute.manager [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 668.300035] env[63531]: DEBUG nova.network.neutron [req-f08a1e88-8879-41e0-9e81-ad5f82f77478 req-911f706d-70f4-42b4-ab05-fe1e5c4a2890 service nova] [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 668.336892] env[63531]: DEBUG nova.network.neutron [-] [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 668.415567] env[63531]: DEBUG nova.network.neutron [req-f08a1e88-8879-41e0-9e81-ad5f82f77478 req-911f706d-70f4-42b4-ab05-fe1e5c4a2890 service nova] [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 668.563169] env[63531]: DEBUG nova.scheduler.client.report [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 668.839909] env[63531]: INFO nova.compute.manager [-] [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] Took 1.02 seconds to deallocate network for instance. [ 668.842821] env[63531]: DEBUG nova.compute.claims [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] Aborting claim: {{(pid=63531) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 668.843014] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 668.918664] env[63531]: DEBUG oslo_concurrency.lockutils [req-f08a1e88-8879-41e0-9e81-ad5f82f77478 req-911f706d-70f4-42b4-ab05-fe1e5c4a2890 service nova] Releasing lock "refresh_cache-9a4f813b-5aac-4458-8fc3-af33b4b9636c" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 668.918664] env[63531]: DEBUG nova.compute.manager [req-f08a1e88-8879-41e0-9e81-ad5f82f77478 req-911f706d-70f4-42b4-ab05-fe1e5c4a2890 service nova] [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] Received event network-vif-deleted-1baeb6e8-6746-4bad-b9da-bce60d8935d5 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 668.989172] env[63531]: ERROR nova.compute.manager [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 50ecfd39-5844-4ad6-b4c9-f29416ae5693, please check neutron logs for more information. [ 668.989172] env[63531]: ERROR nova.compute.manager Traceback (most recent call last): [ 668.989172] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 668.989172] env[63531]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 668.989172] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 668.989172] env[63531]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 668.989172] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 668.989172] env[63531]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 668.989172] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 668.989172] env[63531]: ERROR nova.compute.manager self.force_reraise() [ 668.989172] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 668.989172] env[63531]: ERROR nova.compute.manager raise self.value [ 668.989172] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 668.989172] env[63531]: ERROR nova.compute.manager updated_port = self._update_port( [ 668.989172] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 668.989172] env[63531]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 668.989830] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 668.989830] env[63531]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 668.989830] env[63531]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 50ecfd39-5844-4ad6-b4c9-f29416ae5693, please check neutron logs for more information. [ 668.989830] env[63531]: ERROR nova.compute.manager [ 668.989830] env[63531]: Traceback (most recent call last): [ 668.989830] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 668.989830] env[63531]: listener.cb(fileno) [ 668.989830] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 668.989830] env[63531]: result = function(*args, **kwargs) [ 668.989830] env[63531]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 668.989830] env[63531]: return func(*args, **kwargs) [ 668.989830] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 668.989830] env[63531]: raise e [ 668.989830] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 668.989830] env[63531]: nwinfo = self.network_api.allocate_for_instance( [ 668.989830] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 668.989830] env[63531]: created_port_ids = self._update_ports_for_instance( [ 668.989830] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 668.989830] env[63531]: with excutils.save_and_reraise_exception(): [ 668.989830] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 668.989830] env[63531]: self.force_reraise() [ 668.989830] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 668.989830] env[63531]: raise self.value [ 668.989830] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 668.989830] env[63531]: updated_port = self._update_port( [ 668.989830] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 668.989830] env[63531]: _ensure_no_port_binding_failure(port) [ 668.989830] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 668.989830] env[63531]: raise exception.PortBindingFailed(port_id=port['id']) [ 668.990623] env[63531]: nova.exception.PortBindingFailed: Binding failed for port 50ecfd39-5844-4ad6-b4c9-f29416ae5693, please check neutron logs for more information. [ 668.990623] env[63531]: Removing descriptor: 17 [ 669.068567] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.966s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 669.069217] env[63531]: ERROR nova.compute.manager [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d338839c-7a92-49b7-a6e9-e0abb9e43d33, please check neutron logs for more information. [ 669.069217] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] Traceback (most recent call last): [ 669.069217] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 669.069217] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] self.driver.spawn(context, instance, image_meta, [ 669.069217] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 669.069217] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 669.069217] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 669.069217] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] vm_ref = self.build_virtual_machine(instance, [ 669.069217] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 669.069217] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] vif_infos = vmwarevif.get_vif_info(self._session, [ 669.069217] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 669.069507] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] for vif in network_info: [ 669.069507] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 669.069507] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] return self._sync_wrapper(fn, *args, **kwargs) [ 669.069507] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 669.069507] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] self.wait() [ 669.069507] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 669.069507] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] self[:] = self._gt.wait() [ 669.069507] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 669.069507] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] return self._exit_event.wait() [ 669.069507] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 669.069507] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] result = hub.switch() [ 669.069507] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 669.069507] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] return self.greenlet.switch() [ 669.069928] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 669.069928] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] result = function(*args, **kwargs) [ 669.069928] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 669.069928] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] return func(*args, **kwargs) [ 669.069928] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 669.069928] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] raise e [ 669.069928] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 669.069928] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] nwinfo = self.network_api.allocate_for_instance( [ 669.069928] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 669.069928] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] created_port_ids = self._update_ports_for_instance( [ 669.069928] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 669.069928] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] with excutils.save_and_reraise_exception(): [ 669.069928] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 669.070296] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] self.force_reraise() [ 669.070296] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 669.070296] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] raise self.value [ 669.070296] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 669.070296] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] updated_port = self._update_port( [ 669.070296] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 669.070296] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] _ensure_no_port_binding_failure(port) [ 669.070296] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 669.070296] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] raise exception.PortBindingFailed(port_id=port['id']) [ 669.070296] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] nova.exception.PortBindingFailed: Binding failed for port d338839c-7a92-49b7-a6e9-e0abb9e43d33, please check neutron logs for more information. [ 669.070296] env[63531]: ERROR nova.compute.manager [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] [ 669.070907] env[63531]: DEBUG nova.compute.utils [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] Binding failed for port d338839c-7a92-49b7-a6e9-e0abb9e43d33, please check neutron logs for more information. {{(pid=63531) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 669.071274] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 17.345s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 669.071453] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 669.071615] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63531) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 669.071946] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.555s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 669.074762] env[63531]: DEBUG nova.compute.manager [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] Build of instance e1e72ba0-fc50-4776-991f-738c2d9c4ff3 was re-scheduled: Binding failed for port d338839c-7a92-49b7-a6e9-e0abb9e43d33, please check neutron logs for more information. {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 669.075271] env[63531]: DEBUG nova.compute.manager [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] Unplugging VIFs for instance {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 669.075496] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] Acquiring lock "refresh_cache-e1e72ba0-fc50-4776-991f-738c2d9c4ff3" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 669.075644] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] Acquired lock "refresh_cache-e1e72ba0-fc50-4776-991f-738c2d9c4ff3" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 669.075802] env[63531]: DEBUG nova.network.neutron [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 669.077929] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-346d9eb7-bc51-44f2-aecc-3dd12b3ea234 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.087323] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47cdc3ce-f8e7-4085-94a7-f8081472e2b0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.101778] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a55d7426-ac4e-42cd-baf1-6f3ed1918a12 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.108966] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5c20ce8-edc0-4ccf-852e-10b0b780b416 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.138051] env[63531]: DEBUG nova.compute.manager [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 669.140975] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181510MB free_disk=170GB free_vcpus=48 pci_devices=None {{(pid=63531) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 669.141145] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 669.163556] env[63531]: DEBUG nova.virt.hardware [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 669.163797] env[63531]: DEBUG nova.virt.hardware [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 669.163956] env[63531]: DEBUG nova.virt.hardware [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 669.164157] env[63531]: DEBUG nova.virt.hardware [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 669.164304] env[63531]: DEBUG nova.virt.hardware [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 669.164450] env[63531]: DEBUG nova.virt.hardware [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 669.164657] env[63531]: DEBUG nova.virt.hardware [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 669.164818] env[63531]: DEBUG nova.virt.hardware [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 669.164983] env[63531]: DEBUG nova.virt.hardware [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 669.165163] env[63531]: DEBUG nova.virt.hardware [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 669.165335] env[63531]: DEBUG nova.virt.hardware [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 669.166165] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72057a50-8845-4874-8f55-2a76eb4784aa {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.173824] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eeadd230-6ad9-44ae-a6c7-6cd9ad0b00d2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.188499] env[63531]: ERROR nova.compute.manager [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 50ecfd39-5844-4ad6-b4c9-f29416ae5693, please check neutron logs for more information. [ 669.188499] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] Traceback (most recent call last): [ 669.188499] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 669.188499] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] yield resources [ 669.188499] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 669.188499] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] self.driver.spawn(context, instance, image_meta, [ 669.188499] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 669.188499] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 669.188499] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 669.188499] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] vm_ref = self.build_virtual_machine(instance, [ 669.188499] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 669.188884] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] vif_infos = vmwarevif.get_vif_info(self._session, [ 669.188884] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 669.188884] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] for vif in network_info: [ 669.188884] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 669.188884] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] return self._sync_wrapper(fn, *args, **kwargs) [ 669.188884] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 669.188884] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] self.wait() [ 669.188884] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 669.188884] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] self[:] = self._gt.wait() [ 669.188884] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 669.188884] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] return self._exit_event.wait() [ 669.188884] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 669.188884] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] current.throw(*self._exc) [ 669.189233] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 669.189233] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] result = function(*args, **kwargs) [ 669.189233] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 669.189233] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] return func(*args, **kwargs) [ 669.189233] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 669.189233] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] raise e [ 669.189233] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 669.189233] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] nwinfo = self.network_api.allocate_for_instance( [ 669.189233] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 669.189233] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] created_port_ids = self._update_ports_for_instance( [ 669.189233] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 669.189233] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] with excutils.save_and_reraise_exception(): [ 669.189233] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 669.189609] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] self.force_reraise() [ 669.189609] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 669.189609] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] raise self.value [ 669.189609] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 669.189609] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] updated_port = self._update_port( [ 669.189609] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 669.189609] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] _ensure_no_port_binding_failure(port) [ 669.189609] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 669.189609] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] raise exception.PortBindingFailed(port_id=port['id']) [ 669.189609] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] nova.exception.PortBindingFailed: Binding failed for port 50ecfd39-5844-4ad6-b4c9-f29416ae5693, please check neutron logs for more information. [ 669.189609] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] [ 669.189609] env[63531]: INFO nova.compute.manager [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] Terminating instance [ 669.190756] env[63531]: DEBUG oslo_concurrency.lockutils [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Acquiring lock "refresh_cache-a4dfae30-58f9-4e0b-88b7-c2969d267df6" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 669.190917] env[63531]: DEBUG oslo_concurrency.lockutils [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Acquired lock "refresh_cache-a4dfae30-58f9-4e0b-88b7-c2969d267df6" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 669.191095] env[63531]: DEBUG nova.network.neutron [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 669.600923] env[63531]: DEBUG nova.network.neutron [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 669.672177] env[63531]: DEBUG nova.compute.manager [req-d255389e-148f-4b9b-a865-595da3276e48 req-592df361-23e6-4446-a92f-f1886acd86a6 service nova] [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] Received event network-changed-50ecfd39-5844-4ad6-b4c9-f29416ae5693 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 669.672379] env[63531]: DEBUG nova.compute.manager [req-d255389e-148f-4b9b-a865-595da3276e48 req-592df361-23e6-4446-a92f-f1886acd86a6 service nova] [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] Refreshing instance network info cache due to event network-changed-50ecfd39-5844-4ad6-b4c9-f29416ae5693. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 669.672564] env[63531]: DEBUG oslo_concurrency.lockutils [req-d255389e-148f-4b9b-a865-595da3276e48 req-592df361-23e6-4446-a92f-f1886acd86a6 service nova] Acquiring lock "refresh_cache-a4dfae30-58f9-4e0b-88b7-c2969d267df6" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 669.682622] env[63531]: DEBUG nova.network.neutron [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 669.709042] env[63531]: DEBUG nova.network.neutron [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 669.792035] env[63531]: DEBUG nova.network.neutron [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 669.972810] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d1b57ec-6a58-46e9-a7f6-dcba8699429d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.981305] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d504b97-40e0-4b9b-ac80-7bda28f01e3d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.011466] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27d47029-e016-4195-88d4-93e80bcd7655 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.018874] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1469d3ea-8aa9-4915-90c0-821fb5d20c36 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.031967] env[63531]: DEBUG nova.compute.provider_tree [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 670.185034] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] Releasing lock "refresh_cache-e1e72ba0-fc50-4776-991f-738c2d9c4ff3" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 670.185610] env[63531]: DEBUG nova.compute.manager [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 670.185846] env[63531]: DEBUG nova.compute.manager [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 670.186039] env[63531]: DEBUG nova.network.neutron [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 670.202105] env[63531]: DEBUG nova.network.neutron [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 670.295458] env[63531]: DEBUG oslo_concurrency.lockutils [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Releasing lock "refresh_cache-a4dfae30-58f9-4e0b-88b7-c2969d267df6" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 670.296038] env[63531]: DEBUG nova.compute.manager [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 670.296247] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 670.296629] env[63531]: DEBUG oslo_concurrency.lockutils [req-d255389e-148f-4b9b-a865-595da3276e48 req-592df361-23e6-4446-a92f-f1886acd86a6 service nova] Acquired lock "refresh_cache-a4dfae30-58f9-4e0b-88b7-c2969d267df6" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 670.296800] env[63531]: DEBUG nova.network.neutron [req-d255389e-148f-4b9b-a865-595da3276e48 req-592df361-23e6-4446-a92f-f1886acd86a6 service nova] [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] Refreshing network info cache for port 50ecfd39-5844-4ad6-b4c9-f29416ae5693 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 670.297944] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-90372b2e-34f1-4c30-a231-d42552fa7c32 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.308642] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b03a045-dca1-4f25-9e03-381ffa1c30ac {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.330943] env[63531]: WARNING nova.virt.vmwareapi.vmops [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a4dfae30-58f9-4e0b-88b7-c2969d267df6 could not be found. [ 670.331193] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 670.331378] env[63531]: INFO nova.compute.manager [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] Took 0.04 seconds to destroy the instance on the hypervisor. [ 670.331643] env[63531]: DEBUG oslo.service.loopingcall [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 670.331850] env[63531]: DEBUG nova.compute.manager [-] [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 670.331945] env[63531]: DEBUG nova.network.neutron [-] [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 670.348770] env[63531]: DEBUG nova.network.neutron [-] [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 670.535439] env[63531]: DEBUG nova.scheduler.client.report [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 670.706418] env[63531]: DEBUG nova.network.neutron [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 670.817654] env[63531]: DEBUG nova.network.neutron [req-d255389e-148f-4b9b-a865-595da3276e48 req-592df361-23e6-4446-a92f-f1886acd86a6 service nova] [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 670.851902] env[63531]: DEBUG nova.network.neutron [-] [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 670.892631] env[63531]: DEBUG nova.network.neutron [req-d255389e-148f-4b9b-a865-595da3276e48 req-592df361-23e6-4446-a92f-f1886acd86a6 service nova] [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.040951] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.969s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 671.041627] env[63531]: ERROR nova.compute.manager [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b82524d5-0c39-474f-b157-e64fef4fa205, please check neutron logs for more information. [ 671.041627] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] Traceback (most recent call last): [ 671.041627] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 671.041627] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] self.driver.spawn(context, instance, image_meta, [ 671.041627] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 671.041627] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 671.041627] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 671.041627] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] vm_ref = self.build_virtual_machine(instance, [ 671.041627] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 671.041627] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] vif_infos = vmwarevif.get_vif_info(self._session, [ 671.041627] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 671.041955] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] for vif in network_info: [ 671.041955] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 671.041955] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] return self._sync_wrapper(fn, *args, **kwargs) [ 671.041955] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 671.041955] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] self.wait() [ 671.041955] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 671.041955] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] self[:] = self._gt.wait() [ 671.041955] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 671.041955] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] return self._exit_event.wait() [ 671.041955] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 671.041955] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] result = hub.switch() [ 671.041955] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 671.041955] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] return self.greenlet.switch() [ 671.042269] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 671.042269] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] result = function(*args, **kwargs) [ 671.042269] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 671.042269] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] return func(*args, **kwargs) [ 671.042269] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 671.042269] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] raise e [ 671.042269] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 671.042269] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] nwinfo = self.network_api.allocate_for_instance( [ 671.042269] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 671.042269] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] created_port_ids = self._update_ports_for_instance( [ 671.042269] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 671.042269] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] with excutils.save_and_reraise_exception(): [ 671.042269] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 671.042571] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] self.force_reraise() [ 671.042571] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 671.042571] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] raise self.value [ 671.042571] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 671.042571] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] updated_port = self._update_port( [ 671.042571] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 671.042571] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] _ensure_no_port_binding_failure(port) [ 671.042571] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 671.042571] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] raise exception.PortBindingFailed(port_id=port['id']) [ 671.042571] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] nova.exception.PortBindingFailed: Binding failed for port b82524d5-0c39-474f-b157-e64fef4fa205, please check neutron logs for more information. [ 671.042571] env[63531]: ERROR nova.compute.manager [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] [ 671.042844] env[63531]: DEBUG nova.compute.utils [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] Binding failed for port b82524d5-0c39-474f-b157-e64fef4fa205, please check neutron logs for more information. {{(pid=63531) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 671.043646] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.141s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 671.045292] env[63531]: INFO nova.compute.claims [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 671.049617] env[63531]: DEBUG nova.compute.manager [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] Build of instance 316c12f7-bc05-4f70-98b4-579d4a7b69d0 was re-scheduled: Binding failed for port b82524d5-0c39-474f-b157-e64fef4fa205, please check neutron logs for more information. {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 671.049732] env[63531]: DEBUG nova.compute.manager [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] Unplugging VIFs for instance {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 671.049939] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Acquiring lock "refresh_cache-316c12f7-bc05-4f70-98b4-579d4a7b69d0" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 671.050110] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Acquired lock "refresh_cache-316c12f7-bc05-4f70-98b4-579d4a7b69d0" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 671.050274] env[63531]: DEBUG nova.network.neutron [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 671.209592] env[63531]: INFO nova.compute.manager [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] [instance: e1e72ba0-fc50-4776-991f-738c2d9c4ff3] Took 1.02 seconds to deallocate network for instance. [ 671.358075] env[63531]: INFO nova.compute.manager [-] [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] Took 1.03 seconds to deallocate network for instance. [ 671.360660] env[63531]: DEBUG nova.compute.claims [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] Aborting claim: {{(pid=63531) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 671.360875] env[63531]: DEBUG oslo_concurrency.lockutils [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 671.395761] env[63531]: DEBUG oslo_concurrency.lockutils [req-d255389e-148f-4b9b-a865-595da3276e48 req-592df361-23e6-4446-a92f-f1886acd86a6 service nova] Releasing lock "refresh_cache-a4dfae30-58f9-4e0b-88b7-c2969d267df6" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 671.396130] env[63531]: DEBUG nova.compute.manager [req-d255389e-148f-4b9b-a865-595da3276e48 req-592df361-23e6-4446-a92f-f1886acd86a6 service nova] [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] Received event network-vif-deleted-50ecfd39-5844-4ad6-b4c9-f29416ae5693 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 671.568034] env[63531]: DEBUG nova.network.neutron [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 671.627541] env[63531]: DEBUG nova.network.neutron [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.131745] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Releasing lock "refresh_cache-316c12f7-bc05-4f70-98b4-579d4a7b69d0" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 672.131745] env[63531]: DEBUG nova.compute.manager [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 672.131745] env[63531]: DEBUG nova.compute.manager [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 672.131745] env[63531]: DEBUG nova.network.neutron [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 672.146877] env[63531]: DEBUG nova.network.neutron [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 672.243378] env[63531]: INFO nova.scheduler.client.report [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] Deleted allocations for instance e1e72ba0-fc50-4776-991f-738c2d9c4ff3 [ 672.392186] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71857b1d-32dd-4e1b-8eb4-cf17945fd08b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.399925] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18ebb944-dd7d-4db5-99a5-0216967ea0dc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.429300] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f79e14df-0fba-429a-b66c-aaa079ee145f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.436366] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dcddc14-60ce-41b8-8aba-3231511e6c4d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.449054] env[63531]: DEBUG nova.compute.provider_tree [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 672.649018] env[63531]: DEBUG nova.network.neutron [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.753767] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5792fb0c-6299-4528-8372-e8d6cc70729b tempest-ServersTestManualDisk-23868332 tempest-ServersTestManualDisk-23868332-project-member] Lock "e1e72ba0-fc50-4776-991f-738c2d9c4ff3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 95.857s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 672.952314] env[63531]: DEBUG nova.scheduler.client.report [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 673.153329] env[63531]: INFO nova.compute.manager [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] [instance: 316c12f7-bc05-4f70-98b4-579d4a7b69d0] Took 1.02 seconds to deallocate network for instance. [ 673.256394] env[63531]: DEBUG nova.compute.manager [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: ffb34f90-d22f-440b-ba07-75d474c3c300] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 673.457518] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.414s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 673.458091] env[63531]: DEBUG nova.compute.manager [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 673.460892] env[63531]: DEBUG oslo_concurrency.lockutils [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.686s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 673.787929] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 673.970699] env[63531]: DEBUG nova.compute.utils [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 673.972481] env[63531]: DEBUG nova.compute.manager [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 673.972730] env[63531]: DEBUG nova.network.neutron [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 674.014765] env[63531]: DEBUG nova.policy [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2f0c6adf85584641a2c6d14f56f10674', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ef3b56ce3f3a4155bd09642f5c48ea6f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 674.182023] env[63531]: INFO nova.scheduler.client.report [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Deleted allocations for instance 316c12f7-bc05-4f70-98b4-579d4a7b69d0 [ 674.310120] env[63531]: DEBUG nova.network.neutron [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] Successfully created port: 15f3c89f-b5d4-4d03-8a16-f6522a5af0e0 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 674.329890] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95ebdcca-32a4-42b8-9ee0-176b192ed655 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.338102] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d0c0a05-c4ba-421b-b33e-fb012cd8fb5d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.374111] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d29cbff-a530-4fe8-b958-f9ff910aa266 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.382236] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7a701a5-c632-49d7-aebb-13484b0e2720 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.401213] env[63531]: DEBUG nova.compute.provider_tree [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 674.478607] env[63531]: DEBUG nova.compute.manager [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 674.693843] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6ba7a2fd-c313-489c-b32d-52cd2ed7ab16 tempest-DeleteServersAdminTestJSON-198974442 tempest-DeleteServersAdminTestJSON-198974442-project-member] Lock "316c12f7-bc05-4f70-98b4-579d4a7b69d0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 94.367s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 674.908897] env[63531]: DEBUG nova.scheduler.client.report [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 675.147108] env[63531]: DEBUG nova.compute.manager [req-0785bd39-cafa-4c25-9e7d-9f161c54f24f req-5aaec846-ae55-4350-810c-06cb04b49024 service nova] [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] Received event network-changed-15f3c89f-b5d4-4d03-8a16-f6522a5af0e0 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 675.147108] env[63531]: DEBUG nova.compute.manager [req-0785bd39-cafa-4c25-9e7d-9f161c54f24f req-5aaec846-ae55-4350-810c-06cb04b49024 service nova] [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] Refreshing instance network info cache due to event network-changed-15f3c89f-b5d4-4d03-8a16-f6522a5af0e0. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 675.147108] env[63531]: DEBUG oslo_concurrency.lockutils [req-0785bd39-cafa-4c25-9e7d-9f161c54f24f req-5aaec846-ae55-4350-810c-06cb04b49024 service nova] Acquiring lock "refresh_cache-2a329146-5e0d-4a1f-ba65-3efda2f5e1e0" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 675.147108] env[63531]: DEBUG oslo_concurrency.lockutils [req-0785bd39-cafa-4c25-9e7d-9f161c54f24f req-5aaec846-ae55-4350-810c-06cb04b49024 service nova] Acquired lock "refresh_cache-2a329146-5e0d-4a1f-ba65-3efda2f5e1e0" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 675.151961] env[63531]: DEBUG nova.network.neutron [req-0785bd39-cafa-4c25-9e7d-9f161c54f24f req-5aaec846-ae55-4350-810c-06cb04b49024 service nova] [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] Refreshing network info cache for port 15f3c89f-b5d4-4d03-8a16-f6522a5af0e0 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 675.194953] env[63531]: DEBUG nova.compute.manager [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 675.209735] env[63531]: ERROR nova.compute.manager [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 15f3c89f-b5d4-4d03-8a16-f6522a5af0e0, please check neutron logs for more information. [ 675.209735] env[63531]: ERROR nova.compute.manager Traceback (most recent call last): [ 675.209735] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 675.209735] env[63531]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 675.209735] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 675.209735] env[63531]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 675.209735] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 675.209735] env[63531]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 675.209735] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 675.209735] env[63531]: ERROR nova.compute.manager self.force_reraise() [ 675.209735] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 675.209735] env[63531]: ERROR nova.compute.manager raise self.value [ 675.209735] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 675.209735] env[63531]: ERROR nova.compute.manager updated_port = self._update_port( [ 675.209735] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 675.209735] env[63531]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 675.211019] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 675.211019] env[63531]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 675.211019] env[63531]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 15f3c89f-b5d4-4d03-8a16-f6522a5af0e0, please check neutron logs for more information. [ 675.211019] env[63531]: ERROR nova.compute.manager [ 675.211236] env[63531]: Traceback (most recent call last): [ 675.211236] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 675.211236] env[63531]: listener.cb(fileno) [ 675.211236] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 675.211236] env[63531]: result = function(*args, **kwargs) [ 675.211236] env[63531]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 675.211236] env[63531]: return func(*args, **kwargs) [ 675.211236] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 675.211236] env[63531]: raise e [ 675.211236] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 675.211236] env[63531]: nwinfo = self.network_api.allocate_for_instance( [ 675.211236] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 675.211236] env[63531]: created_port_ids = self._update_ports_for_instance( [ 675.211236] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 675.211236] env[63531]: with excutils.save_and_reraise_exception(): [ 675.211236] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 675.211236] env[63531]: self.force_reraise() [ 675.211236] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 675.211236] env[63531]: raise self.value [ 675.211236] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 675.211236] env[63531]: updated_port = self._update_port( [ 675.211236] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 675.211236] env[63531]: _ensure_no_port_binding_failure(port) [ 675.211236] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 675.211236] env[63531]: raise exception.PortBindingFailed(port_id=port['id']) [ 675.211236] env[63531]: nova.exception.PortBindingFailed: Binding failed for port 15f3c89f-b5d4-4d03-8a16-f6522a5af0e0, please check neutron logs for more information. [ 675.211236] env[63531]: Removing descriptor: 17 [ 675.410160] env[63531]: DEBUG oslo_concurrency.lockutils [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.949s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 675.411484] env[63531]: ERROR nova.compute.manager [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d408a90f-3083-499a-97c9-8090ac70d076, please check neutron logs for more information. [ 675.411484] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] Traceback (most recent call last): [ 675.411484] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 675.411484] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] self.driver.spawn(context, instance, image_meta, [ 675.411484] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 675.411484] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] self._vmops.spawn(context, instance, image_meta, injected_files, [ 675.411484] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 675.411484] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] vm_ref = self.build_virtual_machine(instance, [ 675.411484] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 675.411484] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] vif_infos = vmwarevif.get_vif_info(self._session, [ 675.411484] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 675.411947] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] for vif in network_info: [ 675.411947] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 675.411947] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] return self._sync_wrapper(fn, *args, **kwargs) [ 675.411947] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 675.411947] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] self.wait() [ 675.411947] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 675.411947] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] self[:] = self._gt.wait() [ 675.411947] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 675.411947] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] return self._exit_event.wait() [ 675.411947] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 675.411947] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] current.throw(*self._exc) [ 675.411947] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 675.411947] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] result = function(*args, **kwargs) [ 675.412338] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 675.412338] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] return func(*args, **kwargs) [ 675.412338] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 675.412338] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] raise e [ 675.412338] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 675.412338] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] nwinfo = self.network_api.allocate_for_instance( [ 675.412338] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 675.412338] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] created_port_ids = self._update_ports_for_instance( [ 675.412338] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 675.412338] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] with excutils.save_and_reraise_exception(): [ 675.412338] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 675.412338] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] self.force_reraise() [ 675.412338] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 675.412706] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] raise self.value [ 675.412706] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 675.412706] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] updated_port = self._update_port( [ 675.412706] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 675.412706] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] _ensure_no_port_binding_failure(port) [ 675.412706] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 675.412706] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] raise exception.PortBindingFailed(port_id=port['id']) [ 675.412706] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] nova.exception.PortBindingFailed: Binding failed for port d408a90f-3083-499a-97c9-8090ac70d076, please check neutron logs for more information. [ 675.412706] env[63531]: ERROR nova.compute.manager [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] [ 675.412706] env[63531]: DEBUG nova.compute.utils [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] Binding failed for port d408a90f-3083-499a-97c9-8090ac70d076, please check neutron logs for more information. {{(pid=63531) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 675.414032] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.968s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 675.418159] env[63531]: DEBUG nova.compute.manager [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] Build of instance 180526ef-0e57-4834-bc35-1e8e086e7323 was re-scheduled: Binding failed for port d408a90f-3083-499a-97c9-8090ac70d076, please check neutron logs for more information. {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 675.418159] env[63531]: DEBUG nova.compute.manager [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] Unplugging VIFs for instance {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 675.418159] env[63531]: DEBUG oslo_concurrency.lockutils [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] Acquiring lock "refresh_cache-180526ef-0e57-4834-bc35-1e8e086e7323" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 675.418159] env[63531]: DEBUG oslo_concurrency.lockutils [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] Acquired lock "refresh_cache-180526ef-0e57-4834-bc35-1e8e086e7323" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 675.418329] env[63531]: DEBUG nova.network.neutron [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 675.487030] env[63531]: DEBUG nova.compute.manager [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 675.514304] env[63531]: DEBUG nova.virt.hardware [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 675.514556] env[63531]: DEBUG nova.virt.hardware [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 675.514713] env[63531]: DEBUG nova.virt.hardware [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 675.514898] env[63531]: DEBUG nova.virt.hardware [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 675.515062] env[63531]: DEBUG nova.virt.hardware [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 675.515218] env[63531]: DEBUG nova.virt.hardware [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 675.515424] env[63531]: DEBUG nova.virt.hardware [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 675.515587] env[63531]: DEBUG nova.virt.hardware [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 675.515755] env[63531]: DEBUG nova.virt.hardware [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 675.515918] env[63531]: DEBUG nova.virt.hardware [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 675.516282] env[63531]: DEBUG nova.virt.hardware [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 675.517171] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9206cfd1-da71-4493-bdfe-962f34ec572b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.525524] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20e9491e-c463-4092-8aff-f6e24f776685 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.540455] env[63531]: ERROR nova.compute.manager [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 15f3c89f-b5d4-4d03-8a16-f6522a5af0e0, please check neutron logs for more information. [ 675.540455] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] Traceback (most recent call last): [ 675.540455] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 675.540455] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] yield resources [ 675.540455] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 675.540455] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] self.driver.spawn(context, instance, image_meta, [ 675.540455] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 675.540455] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 675.540455] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 675.540455] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] vm_ref = self.build_virtual_machine(instance, [ 675.540455] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 675.540756] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] vif_infos = vmwarevif.get_vif_info(self._session, [ 675.540756] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 675.540756] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] for vif in network_info: [ 675.540756] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 675.540756] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] return self._sync_wrapper(fn, *args, **kwargs) [ 675.540756] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 675.540756] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] self.wait() [ 675.540756] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 675.540756] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] self[:] = self._gt.wait() [ 675.540756] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 675.540756] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] return self._exit_event.wait() [ 675.540756] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 675.540756] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] current.throw(*self._exc) [ 675.541066] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 675.541066] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] result = function(*args, **kwargs) [ 675.541066] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 675.541066] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] return func(*args, **kwargs) [ 675.541066] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 675.541066] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] raise e [ 675.541066] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 675.541066] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] nwinfo = self.network_api.allocate_for_instance( [ 675.541066] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 675.541066] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] created_port_ids = self._update_ports_for_instance( [ 675.541066] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 675.541066] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] with excutils.save_and_reraise_exception(): [ 675.541066] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 675.541372] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] self.force_reraise() [ 675.541372] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 675.541372] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] raise self.value [ 675.541372] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 675.541372] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] updated_port = self._update_port( [ 675.541372] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 675.541372] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] _ensure_no_port_binding_failure(port) [ 675.541372] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 675.541372] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] raise exception.PortBindingFailed(port_id=port['id']) [ 675.541372] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] nova.exception.PortBindingFailed: Binding failed for port 15f3c89f-b5d4-4d03-8a16-f6522a5af0e0, please check neutron logs for more information. [ 675.541372] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] [ 675.541372] env[63531]: INFO nova.compute.manager [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] Terminating instance [ 675.544799] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Acquiring lock "refresh_cache-2a329146-5e0d-4a1f-ba65-3efda2f5e1e0" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 675.674548] env[63531]: DEBUG nova.network.neutron [req-0785bd39-cafa-4c25-9e7d-9f161c54f24f req-5aaec846-ae55-4350-810c-06cb04b49024 service nova] [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 675.725787] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 675.809404] env[63531]: DEBUG nova.network.neutron [req-0785bd39-cafa-4c25-9e7d-9f161c54f24f req-5aaec846-ae55-4350-810c-06cb04b49024 service nova] [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 675.942549] env[63531]: DEBUG nova.network.neutron [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 676.026591] env[63531]: DEBUG nova.network.neutron [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.312247] env[63531]: DEBUG oslo_concurrency.lockutils [req-0785bd39-cafa-4c25-9e7d-9f161c54f24f req-5aaec846-ae55-4350-810c-06cb04b49024 service nova] Releasing lock "refresh_cache-2a329146-5e0d-4a1f-ba65-3efda2f5e1e0" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 676.315321] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Acquired lock "refresh_cache-2a329146-5e0d-4a1f-ba65-3efda2f5e1e0" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 676.315549] env[63531]: DEBUG nova.network.neutron [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 676.404625] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4d4cfa0-34b2-4246-a122-5410b06f5a81 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.412537] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-394c8ad5-caf6-4035-a7f6-ce4f18e38d81 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.442514] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d5e4b69-d827-41bc-b615-4240394b337b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.449969] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52740725-44ca-4ed5-a05b-b4a8772fe32c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.463915] env[63531]: DEBUG nova.compute.provider_tree [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 676.533033] env[63531]: DEBUG oslo_concurrency.lockutils [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] Releasing lock "refresh_cache-180526ef-0e57-4834-bc35-1e8e086e7323" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 676.533033] env[63531]: DEBUG nova.compute.manager [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 676.533033] env[63531]: DEBUG nova.compute.manager [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 676.533033] env[63531]: DEBUG nova.network.neutron [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 676.547356] env[63531]: DEBUG nova.network.neutron [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 676.834118] env[63531]: DEBUG nova.network.neutron [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 676.907010] env[63531]: DEBUG nova.network.neutron [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.966981] env[63531]: DEBUG nova.scheduler.client.report [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 677.051620] env[63531]: DEBUG nova.network.neutron [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 677.176363] env[63531]: DEBUG nova.compute.manager [req-ae54da07-3897-47de-8535-605168deb01a req-b17e893d-3f8a-4edc-b891-2712f18b84f5 service nova] [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] Received event network-vif-deleted-15f3c89f-b5d4-4d03-8a16-f6522a5af0e0 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 677.409655] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Releasing lock "refresh_cache-2a329146-5e0d-4a1f-ba65-3efda2f5e1e0" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 677.410326] env[63531]: DEBUG nova.compute.manager [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 677.410556] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 677.410905] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-85a41295-03b1-4564-8a91-dcbd62e86fe8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.422985] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb9cbc92-d877-45f9-95bd-1cdb0bcc0104 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.452558] env[63531]: WARNING nova.virt.vmwareapi.vmops [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0 could not be found. [ 677.452788] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 677.453086] env[63531]: INFO nova.compute.manager [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] Took 0.04 seconds to destroy the instance on the hypervisor. [ 677.453350] env[63531]: DEBUG oslo.service.loopingcall [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 677.453583] env[63531]: DEBUG nova.compute.manager [-] [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 677.453682] env[63531]: DEBUG nova.network.neutron [-] [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 677.468212] env[63531]: DEBUG nova.network.neutron [-] [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 677.471528] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.058s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 677.472112] env[63531]: ERROR nova.compute.manager [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 10033ac0-55f0-4735-a120-7e7d5218a6eb, please check neutron logs for more information. [ 677.472112] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] Traceback (most recent call last): [ 677.472112] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 677.472112] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] self.driver.spawn(context, instance, image_meta, [ 677.472112] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 677.472112] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] self._vmops.spawn(context, instance, image_meta, injected_files, [ 677.472112] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 677.472112] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] vm_ref = self.build_virtual_machine(instance, [ 677.472112] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 677.472112] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] vif_infos = vmwarevif.get_vif_info(self._session, [ 677.472112] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 677.472626] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] for vif in network_info: [ 677.472626] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 677.472626] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] return self._sync_wrapper(fn, *args, **kwargs) [ 677.472626] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 677.472626] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] self.wait() [ 677.472626] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 677.472626] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] self[:] = self._gt.wait() [ 677.472626] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 677.472626] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] return self._exit_event.wait() [ 677.472626] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 677.472626] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] current.throw(*self._exc) [ 677.472626] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 677.472626] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] result = function(*args, **kwargs) [ 677.473068] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 677.473068] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] return func(*args, **kwargs) [ 677.473068] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 677.473068] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] raise e [ 677.473068] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 677.473068] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] nwinfo = self.network_api.allocate_for_instance( [ 677.473068] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 677.473068] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] created_port_ids = self._update_ports_for_instance( [ 677.473068] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 677.473068] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] with excutils.save_and_reraise_exception(): [ 677.473068] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 677.473068] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] self.force_reraise() [ 677.473068] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 677.473428] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] raise self.value [ 677.473428] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 677.473428] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] updated_port = self._update_port( [ 677.473428] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 677.473428] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] _ensure_no_port_binding_failure(port) [ 677.473428] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 677.473428] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] raise exception.PortBindingFailed(port_id=port['id']) [ 677.473428] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] nova.exception.PortBindingFailed: Binding failed for port 10033ac0-55f0-4735-a120-7e7d5218a6eb, please check neutron logs for more information. [ 677.473428] env[63531]: ERROR nova.compute.manager [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] [ 677.473428] env[63531]: DEBUG nova.compute.utils [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] Binding failed for port 10033ac0-55f0-4735-a120-7e7d5218a6eb, please check neutron logs for more information. {{(pid=63531) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 677.473908] env[63531]: DEBUG oslo_concurrency.lockutils [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.198s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 677.477481] env[63531]: DEBUG nova.compute.manager [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] Build of instance beb4b334-7d87-4c25-8efe-df92c43f0d26 was re-scheduled: Binding failed for port 10033ac0-55f0-4735-a120-7e7d5218a6eb, please check neutron logs for more information. {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 677.477963] env[63531]: DEBUG nova.compute.manager [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] Unplugging VIFs for instance {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 677.478204] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Acquiring lock "refresh_cache-beb4b334-7d87-4c25-8efe-df92c43f0d26" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 677.478353] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Acquired lock "refresh_cache-beb4b334-7d87-4c25-8efe-df92c43f0d26" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 677.478510] env[63531]: DEBUG nova.network.neutron [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 677.554209] env[63531]: INFO nova.compute.manager [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] [instance: 180526ef-0e57-4834-bc35-1e8e086e7323] Took 1.02 seconds to deallocate network for instance. [ 677.970460] env[63531]: DEBUG nova.network.neutron [-] [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.005934] env[63531]: DEBUG nova.network.neutron [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 678.132731] env[63531]: DEBUG nova.network.neutron [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.359677] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb9caf3f-5fca-49f1-97cd-b2a561b3a05f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.367538] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f93b7342-830b-4287-92e6-5e9238065bba {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.398646] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46ebdc6a-068d-47ec-b431-263fb2ac3b47 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.406564] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38957ae8-3b43-4a8a-b210-7dceddf14e04 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.422896] env[63531]: DEBUG nova.compute.provider_tree [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 678.473210] env[63531]: INFO nova.compute.manager [-] [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] Took 1.02 seconds to deallocate network for instance. [ 678.475894] env[63531]: DEBUG nova.compute.claims [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] Aborting claim: {{(pid=63531) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 678.476096] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 678.593127] env[63531]: INFO nova.scheduler.client.report [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] Deleted allocations for instance 180526ef-0e57-4834-bc35-1e8e086e7323 [ 678.635597] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Releasing lock "refresh_cache-beb4b334-7d87-4c25-8efe-df92c43f0d26" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 678.635873] env[63531]: DEBUG nova.compute.manager [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 678.636152] env[63531]: DEBUG nova.compute.manager [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 678.636332] env[63531]: DEBUG nova.network.neutron [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 678.652914] env[63531]: DEBUG nova.network.neutron [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 678.926921] env[63531]: DEBUG nova.scheduler.client.report [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 679.106711] env[63531]: DEBUG oslo_concurrency.lockutils [None req-80b631d5-bff1-4515-acb6-f53c66a1f735 tempest-InstanceActionsTestJSON-761731010 tempest-InstanceActionsTestJSON-761731010-project-member] Lock "180526ef-0e57-4834-bc35-1e8e086e7323" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 95.144s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 679.157933] env[63531]: DEBUG nova.network.neutron [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 679.434237] env[63531]: DEBUG oslo_concurrency.lockutils [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.960s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 679.435940] env[63531]: ERROR nova.compute.manager [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8b9ee4fb-3390-4145-892c-5750b12ef4e9, please check neutron logs for more information. [ 679.435940] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] Traceback (most recent call last): [ 679.435940] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 679.435940] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] self.driver.spawn(context, instance, image_meta, [ 679.435940] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 679.435940] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 679.435940] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 679.435940] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] vm_ref = self.build_virtual_machine(instance, [ 679.435940] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 679.435940] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] vif_infos = vmwarevif.get_vif_info(self._session, [ 679.435940] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 679.436279] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] for vif in network_info: [ 679.436279] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 679.436279] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] return self._sync_wrapper(fn, *args, **kwargs) [ 679.436279] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 679.436279] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] self.wait() [ 679.436279] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 679.436279] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] self[:] = self._gt.wait() [ 679.436279] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 679.436279] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] return self._exit_event.wait() [ 679.436279] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 679.436279] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] current.throw(*self._exc) [ 679.436279] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 679.436279] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] result = function(*args, **kwargs) [ 679.436629] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 679.436629] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] return func(*args, **kwargs) [ 679.436629] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 679.436629] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] raise e [ 679.436629] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 679.436629] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] nwinfo = self.network_api.allocate_for_instance( [ 679.436629] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 679.436629] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] created_port_ids = self._update_ports_for_instance( [ 679.436629] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 679.436629] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] with excutils.save_and_reraise_exception(): [ 679.436629] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 679.436629] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] self.force_reraise() [ 679.436629] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 679.436928] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] raise self.value [ 679.436928] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 679.436928] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] updated_port = self._update_port( [ 679.436928] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 679.436928] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] _ensure_no_port_binding_failure(port) [ 679.436928] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 679.436928] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] raise exception.PortBindingFailed(port_id=port['id']) [ 679.436928] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] nova.exception.PortBindingFailed: Binding failed for port 8b9ee4fb-3390-4145-892c-5750b12ef4e9, please check neutron logs for more information. [ 679.436928] env[63531]: ERROR nova.compute.manager [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] [ 679.436928] env[63531]: DEBUG nova.compute.utils [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] Binding failed for port 8b9ee4fb-3390-4145-892c-5750b12ef4e9, please check neutron logs for more information. {{(pid=63531) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 679.437531] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.063s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 679.438707] env[63531]: INFO nova.compute.claims [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: 3f3113dd-863f-442d-a214-28dde377d227] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 679.441659] env[63531]: DEBUG nova.compute.manager [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] Build of instance c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f was re-scheduled: Binding failed for port 8b9ee4fb-3390-4145-892c-5750b12ef4e9, please check neutron logs for more information. {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 679.442235] env[63531]: DEBUG nova.compute.manager [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] Unplugging VIFs for instance {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 679.442650] env[63531]: DEBUG oslo_concurrency.lockutils [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] Acquiring lock "refresh_cache-c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 679.442795] env[63531]: DEBUG oslo_concurrency.lockutils [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] Acquired lock "refresh_cache-c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 679.443038] env[63531]: DEBUG nova.network.neutron [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 679.609408] env[63531]: DEBUG nova.compute.manager [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] [instance: 7bc63fdd-4b50-486d-9681-ca0baa08f7b3] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 679.660862] env[63531]: INFO nova.compute.manager [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: beb4b334-7d87-4c25-8efe-df92c43f0d26] Took 1.02 seconds to deallocate network for instance. [ 679.972978] env[63531]: DEBUG nova.network.neutron [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 680.128893] env[63531]: DEBUG nova.network.neutron [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 680.139335] env[63531]: DEBUG oslo_concurrency.lockutils [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 680.632191] env[63531]: DEBUG oslo_concurrency.lockutils [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] Releasing lock "refresh_cache-c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 680.632424] env[63531]: DEBUG nova.compute.manager [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 680.632561] env[63531]: DEBUG nova.compute.manager [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 680.632737] env[63531]: DEBUG nova.network.neutron [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 680.658114] env[63531]: DEBUG nova.network.neutron [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 680.716184] env[63531]: INFO nova.scheduler.client.report [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Deleted allocations for instance beb4b334-7d87-4c25-8efe-df92c43f0d26 [ 680.914034] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6651f9bf-484b-49b3-b228-15add2ee662e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.921698] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ca0fbfe-f313-4215-a2d7-1d24a818fb43 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.954832] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ddc848e-f965-4111-a167-682c77614ad8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.963784] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdfb7257-2f55-461b-ad26-ebce04bf0b83 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.977990] env[63531]: DEBUG nova.compute.provider_tree [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 681.162049] env[63531]: DEBUG nova.network.neutron [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 681.227206] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a0c50a0b-db0e-4150-a52e-9c417f5df35c tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Lock "beb4b334-7d87-4c25-8efe-df92c43f0d26" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 94.637s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 681.485018] env[63531]: DEBUG nova.scheduler.client.report [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 681.664791] env[63531]: INFO nova.compute.manager [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] [instance: c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f] Took 1.03 seconds to deallocate network for instance. [ 681.729336] env[63531]: DEBUG nova.compute.manager [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] [instance: f0276e34-1de3-482a-82c7-1439d40fd85e] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 681.988864] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.552s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 681.990122] env[63531]: DEBUG nova.compute.manager [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: 3f3113dd-863f-442d-a214-28dde377d227] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 681.995524] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.051s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 682.254124] env[63531]: DEBUG oslo_concurrency.lockutils [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 682.501132] env[63531]: DEBUG nova.compute.utils [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 682.505816] env[63531]: DEBUG nova.compute.manager [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: 3f3113dd-863f-442d-a214-28dde377d227] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 682.506391] env[63531]: DEBUG nova.network.neutron [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: 3f3113dd-863f-442d-a214-28dde377d227] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 682.561747] env[63531]: DEBUG nova.policy [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7e9564f4338a42ec8a39aeb679d0e275', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '68629161ef134625aa029648f57223f9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 682.704682] env[63531]: INFO nova.scheduler.client.report [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] Deleted allocations for instance c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f [ 682.930347] env[63531]: DEBUG nova.network.neutron [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: 3f3113dd-863f-442d-a214-28dde377d227] Successfully created port: d0cc21fa-d502-4f4b-99e7-ff4309c06ff5 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 682.964018] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7529582e-8ffe-4c34-8f4d-0f6f01ba7db4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.970142] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e645ed6-0e6d-4d53-a815-60bdd554b5df {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.007868] env[63531]: DEBUG nova.compute.manager [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: 3f3113dd-863f-442d-a214-28dde377d227] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 683.011344] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2326d719-f573-42a4-b424-481ce6719c6c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.019751] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4dd780d-aa7e-4c18-8199-b022fee5b0ac {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.034289] env[63531]: DEBUG nova.compute.provider_tree [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 683.065829] env[63531]: DEBUG oslo_concurrency.lockutils [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Acquiring lock "0197e03d-6c36-4e73-9472-8e3319eb89ce" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 683.066071] env[63531]: DEBUG oslo_concurrency.lockutils [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Lock "0197e03d-6c36-4e73-9472-8e3319eb89ce" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 683.108084] env[63531]: DEBUG oslo_concurrency.lockutils [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Acquiring lock "1e5e28ef-38ce-47be-ac36-8bdf17ab0a47" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 683.108339] env[63531]: DEBUG oslo_concurrency.lockutils [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Lock "1e5e28ef-38ce-47be-ac36-8bdf17ab0a47" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 683.224807] env[63531]: DEBUG oslo_concurrency.lockutils [None req-598b20e6-2807-4cbd-a70c-40328faa1bf6 tempest-ServersTestBootFromVolume-2017991049 tempest-ServersTestBootFromVolume-2017991049-project-member] Lock "c7da97ee-f7a7-4e5e-a567-a8bdee35eb6f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 94.890s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 683.537683] env[63531]: DEBUG nova.scheduler.client.report [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 683.694243] env[63531]: DEBUG nova.compute.manager [req-4097d14b-6787-45f0-9c33-8ee242c7edaa req-42e8a114-98ee-4afc-8cb8-7e2655808eac service nova] [instance: 3f3113dd-863f-442d-a214-28dde377d227] Received event network-changed-d0cc21fa-d502-4f4b-99e7-ff4309c06ff5 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 683.694442] env[63531]: DEBUG nova.compute.manager [req-4097d14b-6787-45f0-9c33-8ee242c7edaa req-42e8a114-98ee-4afc-8cb8-7e2655808eac service nova] [instance: 3f3113dd-863f-442d-a214-28dde377d227] Refreshing instance network info cache due to event network-changed-d0cc21fa-d502-4f4b-99e7-ff4309c06ff5. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 683.694656] env[63531]: DEBUG oslo_concurrency.lockutils [req-4097d14b-6787-45f0-9c33-8ee242c7edaa req-42e8a114-98ee-4afc-8cb8-7e2655808eac service nova] Acquiring lock "refresh_cache-3f3113dd-863f-442d-a214-28dde377d227" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 683.694800] env[63531]: DEBUG oslo_concurrency.lockutils [req-4097d14b-6787-45f0-9c33-8ee242c7edaa req-42e8a114-98ee-4afc-8cb8-7e2655808eac service nova] Acquired lock "refresh_cache-3f3113dd-863f-442d-a214-28dde377d227" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 683.694957] env[63531]: DEBUG nova.network.neutron [req-4097d14b-6787-45f0-9c33-8ee242c7edaa req-42e8a114-98ee-4afc-8cb8-7e2655808eac service nova] [instance: 3f3113dd-863f-442d-a214-28dde377d227] Refreshing network info cache for port d0cc21fa-d502-4f4b-99e7-ff4309c06ff5 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 683.727067] env[63531]: DEBUG nova.compute.manager [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 683.901547] env[63531]: ERROR nova.compute.manager [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d0cc21fa-d502-4f4b-99e7-ff4309c06ff5, please check neutron logs for more information. [ 683.901547] env[63531]: ERROR nova.compute.manager Traceback (most recent call last): [ 683.901547] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 683.901547] env[63531]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 683.901547] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 683.901547] env[63531]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 683.901547] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 683.901547] env[63531]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 683.901547] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 683.901547] env[63531]: ERROR nova.compute.manager self.force_reraise() [ 683.901547] env[63531]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 683.901547] env[63531]: ERROR nova.compute.manager raise self.value [ 683.901547] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 683.901547] env[63531]: ERROR nova.compute.manager updated_port = self._update_port( [ 683.901547] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 683.901547] env[63531]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 683.902083] env[63531]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 683.902083] env[63531]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 683.902083] env[63531]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d0cc21fa-d502-4f4b-99e7-ff4309c06ff5, please check neutron logs for more information. [ 683.902083] env[63531]: ERROR nova.compute.manager [ 683.902083] env[63531]: Traceback (most recent call last): [ 683.902083] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 683.902083] env[63531]: listener.cb(fileno) [ 683.902083] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 683.902083] env[63531]: result = function(*args, **kwargs) [ 683.902083] env[63531]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 683.902083] env[63531]: return func(*args, **kwargs) [ 683.902083] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 683.902083] env[63531]: raise e [ 683.902083] env[63531]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 683.902083] env[63531]: nwinfo = self.network_api.allocate_for_instance( [ 683.902083] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 683.902083] env[63531]: created_port_ids = self._update_ports_for_instance( [ 683.902083] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 683.902083] env[63531]: with excutils.save_and_reraise_exception(): [ 683.902083] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 683.902083] env[63531]: self.force_reraise() [ 683.902083] env[63531]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 683.902083] env[63531]: raise self.value [ 683.902083] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 683.902083] env[63531]: updated_port = self._update_port( [ 683.902083] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 683.902083] env[63531]: _ensure_no_port_binding_failure(port) [ 683.902083] env[63531]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 683.902083] env[63531]: raise exception.PortBindingFailed(port_id=port['id']) [ 683.902904] env[63531]: nova.exception.PortBindingFailed: Binding failed for port d0cc21fa-d502-4f4b-99e7-ff4309c06ff5, please check neutron logs for more information. [ 683.902904] env[63531]: Removing descriptor: 17 [ 684.024891] env[63531]: DEBUG nova.compute.manager [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: 3f3113dd-863f-442d-a214-28dde377d227] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 684.043746] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.050s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 684.046174] env[63531]: ERROR nova.compute.manager [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0f7134ca-5e5e-4a9d-9380-d19042daa7b8, please check neutron logs for more information. [ 684.046174] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] Traceback (most recent call last): [ 684.046174] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 684.046174] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] self.driver.spawn(context, instance, image_meta, [ 684.046174] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 684.046174] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 684.046174] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 684.046174] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] vm_ref = self.build_virtual_machine(instance, [ 684.046174] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 684.046174] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] vif_infos = vmwarevif.get_vif_info(self._session, [ 684.046174] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 684.046484] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] for vif in network_info: [ 684.046484] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 684.046484] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] return self._sync_wrapper(fn, *args, **kwargs) [ 684.046484] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 684.046484] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] self.wait() [ 684.046484] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 684.046484] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] self[:] = self._gt.wait() [ 684.046484] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 684.046484] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] return self._exit_event.wait() [ 684.046484] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 684.046484] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] current.throw(*self._exc) [ 684.046484] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 684.046484] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] result = function(*args, **kwargs) [ 684.046825] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 684.046825] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] return func(*args, **kwargs) [ 684.046825] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 684.046825] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] raise e [ 684.046825] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 684.046825] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] nwinfo = self.network_api.allocate_for_instance( [ 684.046825] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 684.046825] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] created_port_ids = self._update_ports_for_instance( [ 684.046825] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 684.046825] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] with excutils.save_and_reraise_exception(): [ 684.046825] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 684.046825] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] self.force_reraise() [ 684.046825] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 684.047162] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] raise self.value [ 684.047162] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 684.047162] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] updated_port = self._update_port( [ 684.047162] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 684.047162] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] _ensure_no_port_binding_failure(port) [ 684.047162] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 684.047162] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] raise exception.PortBindingFailed(port_id=port['id']) [ 684.047162] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] nova.exception.PortBindingFailed: Binding failed for port 0f7134ca-5e5e-4a9d-9380-d19042daa7b8, please check neutron logs for more information. [ 684.047162] env[63531]: ERROR nova.compute.manager [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] [ 684.047162] env[63531]: DEBUG nova.compute.utils [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] Binding failed for port 0f7134ca-5e5e-4a9d-9380-d19042daa7b8, please check neutron logs for more information. {{(pid=63531) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 684.052166] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.208s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 684.054212] env[63531]: DEBUG nova.compute.manager [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] Build of instance 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f was re-scheduled: Binding failed for port 0f7134ca-5e5e-4a9d-9380-d19042daa7b8, please check neutron logs for more information. {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 684.055381] env[63531]: DEBUG nova.compute.manager [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] Unplugging VIFs for instance {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 684.055381] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] Acquiring lock "refresh_cache-5521c3ab-d017-473d-b8ce-d37c8b2bcc8f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 684.055381] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] Acquired lock "refresh_cache-5521c3ab-d017-473d-b8ce-d37c8b2bcc8f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 684.055381] env[63531]: DEBUG nova.network.neutron [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 684.062097] env[63531]: DEBUG nova.virt.hardware [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 684.062332] env[63531]: DEBUG nova.virt.hardware [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 684.062489] env[63531]: DEBUG nova.virt.hardware [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 684.062674] env[63531]: DEBUG nova.virt.hardware [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 684.062836] env[63531]: DEBUG nova.virt.hardware [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 684.062975] env[63531]: DEBUG nova.virt.hardware [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 684.064470] env[63531]: DEBUG nova.virt.hardware [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 684.064470] env[63531]: DEBUG nova.virt.hardware [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 684.064470] env[63531]: DEBUG nova.virt.hardware [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 684.064470] env[63531]: DEBUG nova.virt.hardware [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 684.064470] env[63531]: DEBUG nova.virt.hardware [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 684.064774] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14d8fec3-eceb-40fb-a686-3974b4e48ee0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.075550] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d95f1c92-1d99-467d-b6da-7058de1d66be {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.093216] env[63531]: ERROR nova.compute.manager [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: 3f3113dd-863f-442d-a214-28dde377d227] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d0cc21fa-d502-4f4b-99e7-ff4309c06ff5, please check neutron logs for more information. [ 684.093216] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] Traceback (most recent call last): [ 684.093216] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 684.093216] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] yield resources [ 684.093216] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 684.093216] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] self.driver.spawn(context, instance, image_meta, [ 684.093216] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 684.093216] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] self._vmops.spawn(context, instance, image_meta, injected_files, [ 684.093216] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 684.093216] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] vm_ref = self.build_virtual_machine(instance, [ 684.093216] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 684.094576] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] vif_infos = vmwarevif.get_vif_info(self._session, [ 684.094576] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 684.094576] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] for vif in network_info: [ 684.094576] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 684.094576] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] return self._sync_wrapper(fn, *args, **kwargs) [ 684.094576] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 684.094576] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] self.wait() [ 684.094576] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 684.094576] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] self[:] = self._gt.wait() [ 684.094576] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 684.094576] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] return self._exit_event.wait() [ 684.094576] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 684.094576] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] current.throw(*self._exc) [ 684.094941] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 684.094941] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] result = function(*args, **kwargs) [ 684.094941] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 684.094941] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] return func(*args, **kwargs) [ 684.094941] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 684.094941] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] raise e [ 684.094941] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 684.094941] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] nwinfo = self.network_api.allocate_for_instance( [ 684.094941] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 684.094941] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] created_port_ids = self._update_ports_for_instance( [ 684.094941] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 684.094941] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] with excutils.save_and_reraise_exception(): [ 684.094941] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 684.095535] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] self.force_reraise() [ 684.095535] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 684.095535] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] raise self.value [ 684.095535] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 684.095535] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] updated_port = self._update_port( [ 684.095535] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 684.095535] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] _ensure_no_port_binding_failure(port) [ 684.095535] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 684.095535] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] raise exception.PortBindingFailed(port_id=port['id']) [ 684.095535] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] nova.exception.PortBindingFailed: Binding failed for port d0cc21fa-d502-4f4b-99e7-ff4309c06ff5, please check neutron logs for more information. [ 684.095535] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] [ 684.095535] env[63531]: INFO nova.compute.manager [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: 3f3113dd-863f-442d-a214-28dde377d227] Terminating instance [ 684.096806] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Acquiring lock "refresh_cache-3f3113dd-863f-442d-a214-28dde377d227" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 684.219604] env[63531]: DEBUG nova.network.neutron [req-4097d14b-6787-45f0-9c33-8ee242c7edaa req-42e8a114-98ee-4afc-8cb8-7e2655808eac service nova] [instance: 3f3113dd-863f-442d-a214-28dde377d227] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 684.268690] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 684.496576] env[63531]: DEBUG nova.network.neutron [req-4097d14b-6787-45f0-9c33-8ee242c7edaa req-42e8a114-98ee-4afc-8cb8-7e2655808eac service nova] [instance: 3f3113dd-863f-442d-a214-28dde377d227] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 684.575495] env[63531]: DEBUG nova.network.neutron [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 684.664213] env[63531]: DEBUG nova.network.neutron [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 684.999860] env[63531]: DEBUG oslo_concurrency.lockutils [req-4097d14b-6787-45f0-9c33-8ee242c7edaa req-42e8a114-98ee-4afc-8cb8-7e2655808eac service nova] Releasing lock "refresh_cache-3f3113dd-863f-442d-a214-28dde377d227" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 685.000300] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Acquired lock "refresh_cache-3f3113dd-863f-442d-a214-28dde377d227" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 685.000483] env[63531]: DEBUG nova.network.neutron [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: 3f3113dd-863f-442d-a214-28dde377d227] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 685.005593] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52640936-f902-411e-8d81-9c10601f8a40 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.015989] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-295ad786-faf7-4a67-8e8d-f961a629447d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.046791] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bc35a43-2b4d-43a0-8c83-ab56ca350323 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.054335] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e58313a-4ea4-48a1-94a0-3f336a49e0b7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.067971] env[63531]: DEBUG nova.compute.provider_tree [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 685.167748] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] Releasing lock "refresh_cache-5521c3ab-d017-473d-b8ce-d37c8b2bcc8f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 685.167748] env[63531]: DEBUG nova.compute.manager [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 685.167748] env[63531]: DEBUG nova.compute.manager [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 685.167748] env[63531]: DEBUG nova.network.neutron [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 685.191589] env[63531]: DEBUG nova.network.neutron [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 685.523231] env[63531]: DEBUG nova.network.neutron [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: 3f3113dd-863f-442d-a214-28dde377d227] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 685.571935] env[63531]: DEBUG nova.scheduler.client.report [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 685.665811] env[63531]: DEBUG nova.network.neutron [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: 3f3113dd-863f-442d-a214-28dde377d227] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 685.695166] env[63531]: DEBUG nova.network.neutron [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 685.715659] env[63531]: DEBUG nova.compute.manager [req-df185976-6ba8-424f-a035-c3f660a54714 req-22017d8f-208d-45cf-80a4-39ce251a9ac0 service nova] [instance: 3f3113dd-863f-442d-a214-28dde377d227] Received event network-vif-deleted-d0cc21fa-d502-4f4b-99e7-ff4309c06ff5 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 686.081258] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.028s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 686.081258] env[63531]: ERROR nova.compute.manager [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1baeb6e8-6746-4bad-b9da-bce60d8935d5, please check neutron logs for more information. [ 686.081258] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] Traceback (most recent call last): [ 686.081258] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 686.081258] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] self.driver.spawn(context, instance, image_meta, [ 686.081258] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 686.081258] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 686.081258] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 686.081258] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] vm_ref = self.build_virtual_machine(instance, [ 686.081750] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 686.081750] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] vif_infos = vmwarevif.get_vif_info(self._session, [ 686.081750] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 686.081750] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] for vif in network_info: [ 686.081750] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 686.081750] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] return self._sync_wrapper(fn, *args, **kwargs) [ 686.081750] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 686.081750] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] self.wait() [ 686.081750] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 686.081750] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] self[:] = self._gt.wait() [ 686.081750] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 686.081750] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] return self._exit_event.wait() [ 686.081750] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 686.082202] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] current.throw(*self._exc) [ 686.082202] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 686.082202] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] result = function(*args, **kwargs) [ 686.082202] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 686.082202] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] return func(*args, **kwargs) [ 686.082202] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 686.082202] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] raise e [ 686.082202] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 686.082202] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] nwinfo = self.network_api.allocate_for_instance( [ 686.082202] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 686.082202] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] created_port_ids = self._update_ports_for_instance( [ 686.082202] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 686.082202] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] with excutils.save_and_reraise_exception(): [ 686.082503] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 686.082503] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] self.force_reraise() [ 686.082503] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 686.082503] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] raise self.value [ 686.082503] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 686.082503] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] updated_port = self._update_port( [ 686.082503] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 686.082503] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] _ensure_no_port_binding_failure(port) [ 686.082503] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 686.082503] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] raise exception.PortBindingFailed(port_id=port['id']) [ 686.082503] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] nova.exception.PortBindingFailed: Binding failed for port 1baeb6e8-6746-4bad-b9da-bce60d8935d5, please check neutron logs for more information. [ 686.082503] env[63531]: ERROR nova.compute.manager [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] [ 686.082790] env[63531]: DEBUG nova.compute.utils [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] Binding failed for port 1baeb6e8-6746-4bad-b9da-bce60d8935d5, please check neutron logs for more information. {{(pid=63531) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 686.082790] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 16.941s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 686.084342] env[63531]: DEBUG nova.compute.manager [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] Build of instance 9a4f813b-5aac-4458-8fc3-af33b4b9636c was re-scheduled: Binding failed for port 1baeb6e8-6746-4bad-b9da-bce60d8935d5, please check neutron logs for more information. {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 686.084850] env[63531]: DEBUG nova.compute.manager [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] Unplugging VIFs for instance {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 686.085143] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] Acquiring lock "refresh_cache-9a4f813b-5aac-4458-8fc3-af33b4b9636c" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 686.085383] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] Acquired lock "refresh_cache-9a4f813b-5aac-4458-8fc3-af33b4b9636c" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 686.085671] env[63531]: DEBUG nova.network.neutron [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 686.171201] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Releasing lock "refresh_cache-3f3113dd-863f-442d-a214-28dde377d227" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 686.171201] env[63531]: DEBUG nova.compute.manager [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: 3f3113dd-863f-442d-a214-28dde377d227] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 686.171321] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: 3f3113dd-863f-442d-a214-28dde377d227] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 686.172029] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bc10f264-4f96-4506-b838-8648357b31f9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.182072] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9130579-7473-43c1-a754-b87ed54a4766 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.198508] env[63531]: INFO nova.compute.manager [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] [instance: 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f] Took 1.03 seconds to deallocate network for instance. [ 686.204842] env[63531]: WARNING nova.virt.vmwareapi.vmops [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: 3f3113dd-863f-442d-a214-28dde377d227] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3f3113dd-863f-442d-a214-28dde377d227 could not be found. [ 686.205099] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: 3f3113dd-863f-442d-a214-28dde377d227] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 686.205290] env[63531]: INFO nova.compute.manager [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: 3f3113dd-863f-442d-a214-28dde377d227] Took 0.03 seconds to destroy the instance on the hypervisor. [ 686.205531] env[63531]: DEBUG oslo.service.loopingcall [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 686.205975] env[63531]: DEBUG nova.compute.manager [-] [instance: 3f3113dd-863f-442d-a214-28dde377d227] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 686.206089] env[63531]: DEBUG nova.network.neutron [-] [instance: 3f3113dd-863f-442d-a214-28dde377d227] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 686.230957] env[63531]: DEBUG nova.network.neutron [-] [instance: 3f3113dd-863f-442d-a214-28dde377d227] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 686.609919] env[63531]: DEBUG nova.network.neutron [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 686.713681] env[63531]: DEBUG nova.network.neutron [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 686.735733] env[63531]: DEBUG nova.network.neutron [-] [instance: 3f3113dd-863f-442d-a214-28dde377d227] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 687.217591] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] Releasing lock "refresh_cache-9a4f813b-5aac-4458-8fc3-af33b4b9636c" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 687.217899] env[63531]: DEBUG nova.compute.manager [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 687.218267] env[63531]: DEBUG nova.compute.manager [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 687.218379] env[63531]: DEBUG nova.network.neutron [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 687.233430] env[63531]: DEBUG nova.network.neutron [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 687.235614] env[63531]: INFO nova.scheduler.client.report [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] Deleted allocations for instance 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f [ 687.241776] env[63531]: INFO nova.compute.manager [-] [instance: 3f3113dd-863f-442d-a214-28dde377d227] Took 1.04 seconds to deallocate network for instance. [ 687.243405] env[63531]: DEBUG nova.network.neutron [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 687.245965] env[63531]: DEBUG nova.compute.claims [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: 3f3113dd-863f-442d-a214-28dde377d227] Aborting claim: {{(pid=63531) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 687.246169] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 687.370700] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquiring lock "9a3fad46-a15a-451c-bdab-a3c8cc8add07" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 687.370967] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "9a3fad46-a15a-451c-bdab-a3c8cc8add07" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 687.616891] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 5521c3ab-d017-473d-b8ce-d37c8b2bcc8f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 687.746813] env[63531]: INFO nova.compute.manager [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] [instance: 9a4f813b-5aac-4458-8fc3-af33b4b9636c] Took 0.53 seconds to deallocate network for instance. [ 687.750060] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62f85dbc-b72f-4edb-940b-537a33e42b9c tempest-ServersNegativeTestMultiTenantJSON-1239999004 tempest-ServersNegativeTestMultiTenantJSON-1239999004-project-member] Lock "5521c3ab-d017-473d-b8ce-d37c8b2bcc8f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 96.951s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 688.120488] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 9a4f813b-5aac-4458-8fc3-af33b4b9636c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 688.120691] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance a4dfae30-58f9-4e0b-88b7-c2969d267df6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 688.120846] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 688.120972] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 3f3113dd-863f-442d-a214-28dde377d227 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 688.252582] env[63531]: DEBUG nova.compute.manager [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: 645abb7a-cf73-4fb0-a9a0-49205f060d6f] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 688.623472] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance ffb34f90-d22f-440b-ba07-75d474c3c300 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 688.785434] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 688.787175] env[63531]: INFO nova.scheduler.client.report [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] Deleted allocations for instance 9a4f813b-5aac-4458-8fc3-af33b4b9636c [ 689.127228] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 0c948655-362d-4673-b3e6-3f28ec69ea3c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 689.294993] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c17dc39b-84b9-495b-80eb-7bc58da20389 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118 tempest-FloatingIPsAssociationNegativeTestJSON-1668158118-project-member] Lock "9a4f813b-5aac-4458-8fc3-af33b4b9636c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 96.361s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 689.631321] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 7bc63fdd-4b50-486d-9681-ca0baa08f7b3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 689.798166] env[63531]: DEBUG nova.compute.manager [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: f39c8807-1e15-41cf-899b-a1fbe0695d58] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 690.135644] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance f0276e34-1de3-482a-82c7-1439d40fd85e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 690.327681] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 690.637764] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance c3d597b4-1dc3-48d8-9bee-e73c8929181b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 691.140579] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 645abb7a-cf73-4fb0-a9a0-49205f060d6f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 691.643858] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance f39c8807-1e15-41cf-899b-a1fbe0695d58 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 692.147220] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 5e9042b3-4dc2-4fa3-a664-c4b49a22e400 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 692.651963] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance c3920e60-b27f-496b-b616-afb23ab0bc26 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 693.201932] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 410ddf7a-e45b-4769-a3db-5363ce2096a7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 693.663687] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 1ca7b851-2a06-4181-8271-58aafcd322d6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 694.166936] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance e7fb00a3-0d2e-4f54-950d-337307112d7a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 694.670418] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 0af99240-fa7a-4eed-a729-a5ae98c41cf9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 695.175316] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 695.678178] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance e1693594-d8bb-4002-b71c-7013efb81d16 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 695.898236] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquiring lock "48043b5d-c0d3-4e8a-8d28-5a408d800e11" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 695.898469] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "48043b5d-c0d3-4e8a-8d28-5a408d800e11" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 696.183762] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 29dd6de3-2765-408c-acc0-da47e5e0a977 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 696.685366] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance dccb2335-4220-4570-861b-16ff2d8f5a85 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 697.188364] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 8dec0cac-defc-42f2-ab56-6b3ae60ad858 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 697.692705] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 698.196495] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 64fb5831-0789-482b-bb98-67b29868c4c9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 698.698714] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 0197e03d-6c36-4e73-9472-8e3319eb89ce has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 699.202295] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 699.706669] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 9a3fad46-a15a-451c-bdab-a3c8cc8add07 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 699.707017] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=63531) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 699.707107] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=63531) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 700.003744] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bc1a745-3e61-4377-95b3-702b24604783 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.012173] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33863510-34d9-45cc-8c4e-b79cd0b34f0c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.041824] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eacc6ff1-2732-40bc-9b0c-edb3b175b32c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.049034] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ceb3e7a-9a5f-4e56-afc5-67be5baefab4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.062634] env[63531]: DEBUG nova.compute.provider_tree [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 700.565645] env[63531]: DEBUG nova.scheduler.client.report [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 701.071468] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63531) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 701.071741] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 14.989s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 701.072096] env[63531]: DEBUG oslo_concurrency.lockutils [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.711s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 701.845076] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78d59999-3b4c-4dc2-aa03-afb4c3d89688 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.852701] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4deb1c68-3cdb-4da6-97c5-aa6215a9d377 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.881916] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ec8fd28-832f-45ee-9726-66d1a4e162c2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.889547] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29b8398c-f99a-48de-9689-f6d199034dcb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.902229] env[63531]: DEBUG nova.compute.provider_tree [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 702.407611] env[63531]: DEBUG nova.scheduler.client.report [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 702.915284] env[63531]: DEBUG oslo_concurrency.lockutils [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.843s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 702.915789] env[63531]: ERROR nova.compute.manager [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 50ecfd39-5844-4ad6-b4c9-f29416ae5693, please check neutron logs for more information. [ 702.915789] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] Traceback (most recent call last): [ 702.915789] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 702.915789] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] self.driver.spawn(context, instance, image_meta, [ 702.915789] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 702.915789] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 702.915789] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 702.915789] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] vm_ref = self.build_virtual_machine(instance, [ 702.915789] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 702.915789] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] vif_infos = vmwarevif.get_vif_info(self._session, [ 702.915789] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 702.916143] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] for vif in network_info: [ 702.916143] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 702.916143] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] return self._sync_wrapper(fn, *args, **kwargs) [ 702.916143] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 702.916143] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] self.wait() [ 702.916143] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 702.916143] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] self[:] = self._gt.wait() [ 702.916143] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 702.916143] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] return self._exit_event.wait() [ 702.916143] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 702.916143] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] current.throw(*self._exc) [ 702.916143] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 702.916143] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] result = function(*args, **kwargs) [ 702.916627] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 702.916627] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] return func(*args, **kwargs) [ 702.916627] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 702.916627] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] raise e [ 702.916627] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 702.916627] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] nwinfo = self.network_api.allocate_for_instance( [ 702.916627] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 702.916627] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] created_port_ids = self._update_ports_for_instance( [ 702.916627] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 702.916627] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] with excutils.save_and_reraise_exception(): [ 702.916627] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 702.916627] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] self.force_reraise() [ 702.916627] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 702.916939] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] raise self.value [ 702.916939] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 702.916939] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] updated_port = self._update_port( [ 702.916939] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 702.916939] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] _ensure_no_port_binding_failure(port) [ 702.916939] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 702.916939] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] raise exception.PortBindingFailed(port_id=port['id']) [ 702.916939] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] nova.exception.PortBindingFailed: Binding failed for port 50ecfd39-5844-4ad6-b4c9-f29416ae5693, please check neutron logs for more information. [ 702.916939] env[63531]: ERROR nova.compute.manager [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] [ 702.916939] env[63531]: DEBUG nova.compute.utils [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] Binding failed for port 50ecfd39-5844-4ad6-b4c9-f29416ae5693, please check neutron logs for more information. {{(pid=63531) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 702.917822] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.130s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 702.919278] env[63531]: INFO nova.compute.claims [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: ffb34f90-d22f-440b-ba07-75d474c3c300] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 702.922042] env[63531]: DEBUG nova.compute.manager [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] Build of instance a4dfae30-58f9-4e0b-88b7-c2969d267df6 was re-scheduled: Binding failed for port 50ecfd39-5844-4ad6-b4c9-f29416ae5693, please check neutron logs for more information. {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 702.922182] env[63531]: DEBUG nova.compute.manager [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] Unplugging VIFs for instance {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 702.922406] env[63531]: DEBUG oslo_concurrency.lockutils [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Acquiring lock "refresh_cache-a4dfae30-58f9-4e0b-88b7-c2969d267df6" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 702.922549] env[63531]: DEBUG oslo_concurrency.lockutils [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Acquired lock "refresh_cache-a4dfae30-58f9-4e0b-88b7-c2969d267df6" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 702.922714] env[63531]: DEBUG nova.network.neutron [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 703.443502] env[63531]: DEBUG nova.network.neutron [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 703.521594] env[63531]: DEBUG nova.network.neutron [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.024478] env[63531]: DEBUG oslo_concurrency.lockutils [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Releasing lock "refresh_cache-a4dfae30-58f9-4e0b-88b7-c2969d267df6" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 704.024714] env[63531]: DEBUG nova.compute.manager [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 704.024906] env[63531]: DEBUG nova.compute.manager [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 704.025085] env[63531]: DEBUG nova.network.neutron [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 704.040861] env[63531]: DEBUG nova.network.neutron [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 704.213453] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a828dc0-83c0-4f71-bfa0-7757741863d1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.221042] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edb22160-5e62-4172-b7dc-e687c9d4dd5c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.250107] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18f9a689-1a2f-4a2d-b0c7-3cb89982e02d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.257360] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea688713-e6f8-456d-8afb-4e358f0ae883 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.270256] env[63531]: DEBUG nova.compute.provider_tree [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 704.543209] env[63531]: DEBUG nova.network.neutron [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.773204] env[63531]: DEBUG nova.scheduler.client.report [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 705.050264] env[63531]: INFO nova.compute.manager [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: a4dfae30-58f9-4e0b-88b7-c2969d267df6] Took 1.02 seconds to deallocate network for instance. [ 705.278560] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.361s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 705.279137] env[63531]: DEBUG nova.compute.manager [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: ffb34f90-d22f-440b-ba07-75d474c3c300] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 705.283138] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.556s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 705.283537] env[63531]: INFO nova.compute.claims [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 705.790594] env[63531]: DEBUG nova.compute.utils [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 705.790866] env[63531]: DEBUG nova.compute.manager [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: ffb34f90-d22f-440b-ba07-75d474c3c300] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 705.791013] env[63531]: DEBUG nova.network.neutron [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: ffb34f90-d22f-440b-ba07-75d474c3c300] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 705.842863] env[63531]: DEBUG nova.policy [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7e9564f4338a42ec8a39aeb679d0e275', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '68629161ef134625aa029648f57223f9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 706.082056] env[63531]: INFO nova.scheduler.client.report [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Deleted allocations for instance a4dfae30-58f9-4e0b-88b7-c2969d267df6 [ 706.090331] env[63531]: DEBUG nova.network.neutron [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: ffb34f90-d22f-440b-ba07-75d474c3c300] Successfully created port: 17d053f9-c98c-4b3d-8a11-1c514de74d81 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 706.295201] env[63531]: DEBUG nova.compute.manager [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: ffb34f90-d22f-440b-ba07-75d474c3c300] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 706.588631] env[63531]: DEBUG oslo_concurrency.lockutils [None req-2ce791d3-80b3-479c-af67-d7ad217e3531 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Lock "a4dfae30-58f9-4e0b-88b7-c2969d267df6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 112.388s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 706.619713] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a24dbb9c-2b8c-4c56-9a69-95590579ccb1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.627168] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc88e236-020c-44be-a2de-2bc6f62f0349 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.660132] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35d96982-53b2-420a-9b89-8bfea9b45081 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.667458] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-242b9b22-0e17-41f2-bd2c-e58bd10b27ce {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.681074] env[63531]: DEBUG nova.compute.provider_tree [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 707.091865] env[63531]: DEBUG nova.compute.manager [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 707.183414] env[63531]: DEBUG nova.scheduler.client.report [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 707.306472] env[63531]: DEBUG nova.compute.manager [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: ffb34f90-d22f-440b-ba07-75d474c3c300] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 707.332585] env[63531]: DEBUG nova.virt.hardware [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 707.332979] env[63531]: DEBUG nova.virt.hardware [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 707.333274] env[63531]: DEBUG nova.virt.hardware [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 707.333579] env[63531]: DEBUG nova.virt.hardware [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 707.333829] env[63531]: DEBUG nova.virt.hardware [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 707.334090] env[63531]: DEBUG nova.virt.hardware [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 707.334444] env[63531]: DEBUG nova.virt.hardware [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 707.334735] env[63531]: DEBUG nova.virt.hardware [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 707.334990] env[63531]: DEBUG nova.virt.hardware [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 707.335288] env[63531]: DEBUG nova.virt.hardware [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 707.335598] env[63531]: DEBUG nova.virt.hardware [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 707.336952] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfcd4438-1d7e-4746-a64c-f1d6ba5d1cab {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.348779] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd83004f-cc7b-437f-9c09-f6e186f250be {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.566175] env[63531]: DEBUG nova.compute.manager [req-536c17e1-166d-4623-aa83-d398d956c8d2 req-3b1af4aa-67e4-4fe5-b0b3-17b05c7e70c3 service nova] [instance: ffb34f90-d22f-440b-ba07-75d474c3c300] Received event network-vif-plugged-17d053f9-c98c-4b3d-8a11-1c514de74d81 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 707.566406] env[63531]: DEBUG oslo_concurrency.lockutils [req-536c17e1-166d-4623-aa83-d398d956c8d2 req-3b1af4aa-67e4-4fe5-b0b3-17b05c7e70c3 service nova] Acquiring lock "ffb34f90-d22f-440b-ba07-75d474c3c300-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 707.566680] env[63531]: DEBUG oslo_concurrency.lockutils [req-536c17e1-166d-4623-aa83-d398d956c8d2 req-3b1af4aa-67e4-4fe5-b0b3-17b05c7e70c3 service nova] Lock "ffb34f90-d22f-440b-ba07-75d474c3c300-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 707.567010] env[63531]: DEBUG oslo_concurrency.lockutils [req-536c17e1-166d-4623-aa83-d398d956c8d2 req-3b1af4aa-67e4-4fe5-b0b3-17b05c7e70c3 service nova] Lock "ffb34f90-d22f-440b-ba07-75d474c3c300-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 707.567099] env[63531]: DEBUG nova.compute.manager [req-536c17e1-166d-4623-aa83-d398d956c8d2 req-3b1af4aa-67e4-4fe5-b0b3-17b05c7e70c3 service nova] [instance: ffb34f90-d22f-440b-ba07-75d474c3c300] No waiting events found dispatching network-vif-plugged-17d053f9-c98c-4b3d-8a11-1c514de74d81 {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 707.567208] env[63531]: WARNING nova.compute.manager [req-536c17e1-166d-4623-aa83-d398d956c8d2 req-3b1af4aa-67e4-4fe5-b0b3-17b05c7e70c3 service nova] [instance: ffb34f90-d22f-440b-ba07-75d474c3c300] Received unexpected event network-vif-plugged-17d053f9-c98c-4b3d-8a11-1c514de74d81 for instance with vm_state building and task_state spawning. [ 707.620322] env[63531]: DEBUG nova.network.neutron [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: ffb34f90-d22f-440b-ba07-75d474c3c300] Successfully updated port: 17d053f9-c98c-4b3d-8a11-1c514de74d81 {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 707.622370] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 707.689395] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.408s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 707.690015] env[63531]: DEBUG nova.compute.manager [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 707.696107] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.217s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 708.124499] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Acquiring lock "refresh_cache-ffb34f90-d22f-440b-ba07-75d474c3c300" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 708.124753] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Acquired lock "refresh_cache-ffb34f90-d22f-440b-ba07-75d474c3c300" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 708.124850] env[63531]: DEBUG nova.network.neutron [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: ffb34f90-d22f-440b-ba07-75d474c3c300] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 708.202185] env[63531]: DEBUG nova.compute.utils [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 708.203916] env[63531]: DEBUG nova.compute.manager [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Not allocating networking since 'none' was specified. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 708.532288] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6d2195c-a822-458e-b847-19a6f21d9700 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.540440] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22eebb7e-719a-4bff-96d8-59bcef8db8fc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.570198] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c1c61e4-4601-40d6-aa70-40a5d7195e08 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.577445] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd88fd8b-9712-4ad6-a5a6-12cd3aaef9f3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.590262] env[63531]: DEBUG nova.compute.provider_tree [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 708.667939] env[63531]: DEBUG nova.network.neutron [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: ffb34f90-d22f-440b-ba07-75d474c3c300] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 708.706057] env[63531]: DEBUG nova.compute.manager [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 708.857090] env[63531]: DEBUG nova.network.neutron [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: ffb34f90-d22f-440b-ba07-75d474c3c300] Updating instance_info_cache with network_info: [{"id": "17d053f9-c98c-4b3d-8a11-1c514de74d81", "address": "fa:16:3e:ed:f2:4a", "network": {"id": "c3fa0230-dd92-4cd1-a79b-b7aa41837058", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-590933268-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "68629161ef134625aa029648f57223f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "79c2e589-f55b-4843-8d99-2e565be16706", "external-id": "nsx-vlan-transportzone-858", "segmentation_id": 858, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap17d053f9-c9", "ovs_interfaceid": "17d053f9-c98c-4b3d-8a11-1c514de74d81", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 709.093229] env[63531]: DEBUG nova.scheduler.client.report [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 709.359869] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Releasing lock "refresh_cache-ffb34f90-d22f-440b-ba07-75d474c3c300" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 709.360242] env[63531]: DEBUG nova.compute.manager [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: ffb34f90-d22f-440b-ba07-75d474c3c300] Instance network_info: |[{"id": "17d053f9-c98c-4b3d-8a11-1c514de74d81", "address": "fa:16:3e:ed:f2:4a", "network": {"id": "c3fa0230-dd92-4cd1-a79b-b7aa41837058", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-590933268-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "68629161ef134625aa029648f57223f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "79c2e589-f55b-4843-8d99-2e565be16706", "external-id": "nsx-vlan-transportzone-858", "segmentation_id": 858, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap17d053f9-c9", "ovs_interfaceid": "17d053f9-c98c-4b3d-8a11-1c514de74d81", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 709.360689] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: ffb34f90-d22f-440b-ba07-75d474c3c300] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ed:f2:4a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '79c2e589-f55b-4843-8d99-2e565be16706', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '17d053f9-c98c-4b3d-8a11-1c514de74d81', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 709.368574] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Creating folder: Project (68629161ef134625aa029648f57223f9). Parent ref: group-v244585. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 709.368863] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e6ee48ef-1d85-42e9-a358-227f323f93f2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.381561] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Created folder: Project (68629161ef134625aa029648f57223f9) in parent group-v244585. [ 709.381749] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Creating folder: Instances. Parent ref: group-v244602. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 709.381970] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ef355ac5-4929-4de4-8a6f-6b2f912bcb6c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.392273] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Created folder: Instances in parent group-v244602. [ 709.392490] env[63531]: DEBUG oslo.service.loopingcall [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 709.392673] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ffb34f90-d22f-440b-ba07-75d474c3c300] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 709.392865] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f4a51b02-1a16-434d-b0b3-125ff6e94b37 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.410417] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 709.410417] env[63531]: value = "task-1117531" [ 709.410417] env[63531]: _type = "Task" [ 709.410417] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.418721] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117531, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.594906] env[63531]: DEBUG nova.compute.manager [req-b6e3d250-cd92-48fb-b457-4253ce852a14 req-e61f762d-af4c-4452-a7b8-59995bb835c3 service nova] [instance: ffb34f90-d22f-440b-ba07-75d474c3c300] Received event network-changed-17d053f9-c98c-4b3d-8a11-1c514de74d81 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 709.595256] env[63531]: DEBUG nova.compute.manager [req-b6e3d250-cd92-48fb-b457-4253ce852a14 req-e61f762d-af4c-4452-a7b8-59995bb835c3 service nova] [instance: ffb34f90-d22f-440b-ba07-75d474c3c300] Refreshing instance network info cache due to event network-changed-17d053f9-c98c-4b3d-8a11-1c514de74d81. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 709.595595] env[63531]: DEBUG oslo_concurrency.lockutils [req-b6e3d250-cd92-48fb-b457-4253ce852a14 req-e61f762d-af4c-4452-a7b8-59995bb835c3 service nova] Acquiring lock "refresh_cache-ffb34f90-d22f-440b-ba07-75d474c3c300" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 709.595762] env[63531]: DEBUG oslo_concurrency.lockutils [req-b6e3d250-cd92-48fb-b457-4253ce852a14 req-e61f762d-af4c-4452-a7b8-59995bb835c3 service nova] Acquired lock "refresh_cache-ffb34f90-d22f-440b-ba07-75d474c3c300" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 709.595943] env[63531]: DEBUG nova.network.neutron [req-b6e3d250-cd92-48fb-b457-4253ce852a14 req-e61f762d-af4c-4452-a7b8-59995bb835c3 service nova] [instance: ffb34f90-d22f-440b-ba07-75d474c3c300] Refreshing network info cache for port 17d053f9-c98c-4b3d-8a11-1c514de74d81 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 709.598995] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.905s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 709.598995] env[63531]: ERROR nova.compute.manager [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 15f3c89f-b5d4-4d03-8a16-f6522a5af0e0, please check neutron logs for more information. [ 709.598995] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] Traceback (most recent call last): [ 709.598995] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 709.598995] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] self.driver.spawn(context, instance, image_meta, [ 709.598995] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 709.598995] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 709.598995] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 709.598995] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] vm_ref = self.build_virtual_machine(instance, [ 709.599299] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 709.599299] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] vif_infos = vmwarevif.get_vif_info(self._session, [ 709.599299] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 709.599299] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] for vif in network_info: [ 709.599299] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 709.599299] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] return self._sync_wrapper(fn, *args, **kwargs) [ 709.599299] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 709.599299] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] self.wait() [ 709.599299] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 709.599299] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] self[:] = self._gt.wait() [ 709.599299] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 709.599299] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] return self._exit_event.wait() [ 709.599299] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 709.599670] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] current.throw(*self._exc) [ 709.599670] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 709.599670] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] result = function(*args, **kwargs) [ 709.599670] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 709.599670] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] return func(*args, **kwargs) [ 709.599670] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 709.599670] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] raise e [ 709.599670] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 709.599670] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] nwinfo = self.network_api.allocate_for_instance( [ 709.599670] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 709.599670] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] created_port_ids = self._update_ports_for_instance( [ 709.599670] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 709.599670] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] with excutils.save_and_reraise_exception(): [ 709.599988] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 709.599988] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] self.force_reraise() [ 709.599988] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 709.599988] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] raise self.value [ 709.599988] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 709.599988] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] updated_port = self._update_port( [ 709.599988] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 709.599988] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] _ensure_no_port_binding_failure(port) [ 709.599988] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 709.599988] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] raise exception.PortBindingFailed(port_id=port['id']) [ 709.599988] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] nova.exception.PortBindingFailed: Binding failed for port 15f3c89f-b5d4-4d03-8a16-f6522a5af0e0, please check neutron logs for more information. [ 709.599988] env[63531]: ERROR nova.compute.manager [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] [ 709.600296] env[63531]: DEBUG nova.compute.utils [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] Binding failed for port 15f3c89f-b5d4-4d03-8a16-f6522a5af0e0, please check neutron logs for more information. {{(pid=63531) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 709.601096] env[63531]: DEBUG nova.compute.manager [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] Build of instance 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0 was re-scheduled: Binding failed for port 15f3c89f-b5d4-4d03-8a16-f6522a5af0e0, please check neutron logs for more information. {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 709.601518] env[63531]: DEBUG nova.compute.manager [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] Unplugging VIFs for instance {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 709.601731] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Acquiring lock "refresh_cache-2a329146-5e0d-4a1f-ba65-3efda2f5e1e0" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 709.601878] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Acquired lock "refresh_cache-2a329146-5e0d-4a1f-ba65-3efda2f5e1e0" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 709.602062] env[63531]: DEBUG nova.network.neutron [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 709.603568] env[63531]: DEBUG oslo_concurrency.lockutils [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.464s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 709.605015] env[63531]: INFO nova.compute.claims [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] [instance: 7bc63fdd-4b50-486d-9681-ca0baa08f7b3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 709.642751] env[63531]: DEBUG oslo_concurrency.lockutils [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Acquiring lock "caf0dee8-cb5a-402b-b533-22a590433f54" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 709.642751] env[63531]: DEBUG oslo_concurrency.lockutils [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Lock "caf0dee8-cb5a-402b-b533-22a590433f54" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 709.715869] env[63531]: DEBUG nova.compute.manager [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 709.743145] env[63531]: DEBUG nova.virt.hardware [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 709.743398] env[63531]: DEBUG nova.virt.hardware [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 709.743546] env[63531]: DEBUG nova.virt.hardware [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 709.744150] env[63531]: DEBUG nova.virt.hardware [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 709.744150] env[63531]: DEBUG nova.virt.hardware [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 709.744150] env[63531]: DEBUG nova.virt.hardware [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 709.744304] env[63531]: DEBUG nova.virt.hardware [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 709.744371] env[63531]: DEBUG nova.virt.hardware [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 709.744610] env[63531]: DEBUG nova.virt.hardware [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 709.744762] env[63531]: DEBUG nova.virt.hardware [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 709.744943] env[63531]: DEBUG nova.virt.hardware [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 709.745794] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-860bc6ea-bb9a-4d9e-9f83-015159b46282 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.753829] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f4744b1-67cd-42a5-aafe-688b09c2f52d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.767971] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Instance VIF info [] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 709.772584] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Creating folder: Project (a24db89948554390a1a3353354441c11). Parent ref: group-v244585. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 709.773203] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7713d090-ba9b-44b4-a991-d3402cd01e35 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.783608] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Created folder: Project (a24db89948554390a1a3353354441c11) in parent group-v244585. [ 709.783802] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Creating folder: Instances. Parent ref: group-v244605. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 709.784040] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9fe7889a-9cf6-453f-9336-43277e86d3cf {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.791859] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Created folder: Instances in parent group-v244605. [ 709.791960] env[63531]: DEBUG oslo.service.loopingcall [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 709.792173] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 709.792365] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f3cf92b9-8490-4e7f-99ce-f5d0b205ebf2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.808133] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 709.808133] env[63531]: value = "task-1117534" [ 709.808133] env[63531]: _type = "Task" [ 709.808133] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.815056] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117534, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.920079] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117531, 'name': CreateVM_Task, 'duration_secs': 0.311184} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.920260] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ffb34f90-d22f-440b-ba07-75d474c3c300] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 709.927043] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 709.927216] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 709.927544] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 709.927811] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-722b94db-fcda-4b39-b0e2-baf3c8c6fdcd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.932300] env[63531]: DEBUG oslo_vmware.api [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Waiting for the task: (returnval){ [ 709.932300] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52c5214d-9d8e-f96f-bdd3-6e123d6ecb61" [ 709.932300] env[63531]: _type = "Task" [ 709.932300] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.940154] env[63531]: DEBUG oslo_vmware.api [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52c5214d-9d8e-f96f-bdd3-6e123d6ecb61, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.122498] env[63531]: DEBUG nova.network.neutron [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 710.318695] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117534, 'name': CreateVM_Task, 'duration_secs': 0.271675} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.318865] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 710.319288] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 710.393474] env[63531]: DEBUG nova.network.neutron [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 710.442692] env[63531]: DEBUG oslo_vmware.api [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52c5214d-9d8e-f96f-bdd3-6e123d6ecb61, 'name': SearchDatastore_Task, 'duration_secs': 0.012296} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.443017] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 710.443230] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: ffb34f90-d22f-440b-ba07-75d474c3c300] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 710.443502] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 710.443660] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 710.443842] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 710.444446] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 710.444446] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 710.444635] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6a1d0861-1aff-4a25-af4f-e45b45baf971 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.447227] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a6bac1fa-1886-4258-8050-e7cb7eeda052 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.454952] env[63531]: DEBUG oslo_vmware.api [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Waiting for the task: (returnval){ [ 710.454952] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52c5c83b-f9c1-ddbc-bc5a-d738a78e7ccc" [ 710.454952] env[63531]: _type = "Task" [ 710.454952] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.456089] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 710.456268] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 710.459620] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bb43f1fd-3492-4f30-9d39-75caa286ce47 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.469569] env[63531]: DEBUG oslo_vmware.api [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52c5c83b-f9c1-ddbc-bc5a-d738a78e7ccc, 'name': SearchDatastore_Task, 'duration_secs': 0.009592} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.473050] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 710.473296] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 710.473505] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 710.473774] env[63531]: DEBUG oslo_vmware.api [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Waiting for the task: (returnval){ [ 710.473774] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52cfea36-87ec-ac5c-7b7a-5626d9d48bff" [ 710.473774] env[63531]: _type = "Task" [ 710.473774] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.482511] env[63531]: DEBUG oslo_vmware.api [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52cfea36-87ec-ac5c-7b7a-5626d9d48bff, 'name': SearchDatastore_Task, 'duration_secs': 0.009905} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.483486] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-423f10f0-96ce-4597-92ba-5b29cd4c8064 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.488199] env[63531]: DEBUG oslo_vmware.api [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Waiting for the task: (returnval){ [ 710.488199] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5257dfc3-7b81-a5ad-05f1-53a58964b37c" [ 710.488199] env[63531]: _type = "Task" [ 710.488199] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.495294] env[63531]: DEBUG oslo_vmware.api [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5257dfc3-7b81-a5ad-05f1-53a58964b37c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.519153] env[63531]: DEBUG nova.network.neutron [req-b6e3d250-cd92-48fb-b457-4253ce852a14 req-e61f762d-af4c-4452-a7b8-59995bb835c3 service nova] [instance: ffb34f90-d22f-440b-ba07-75d474c3c300] Updated VIF entry in instance network info cache for port 17d053f9-c98c-4b3d-8a11-1c514de74d81. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 710.519512] env[63531]: DEBUG nova.network.neutron [req-b6e3d250-cd92-48fb-b457-4253ce852a14 req-e61f762d-af4c-4452-a7b8-59995bb835c3 service nova] [instance: ffb34f90-d22f-440b-ba07-75d474c3c300] Updating instance_info_cache with network_info: [{"id": "17d053f9-c98c-4b3d-8a11-1c514de74d81", "address": "fa:16:3e:ed:f2:4a", "network": {"id": "c3fa0230-dd92-4cd1-a79b-b7aa41837058", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-590933268-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "68629161ef134625aa029648f57223f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "79c2e589-f55b-4843-8d99-2e565be16706", "external-id": "nsx-vlan-transportzone-858", "segmentation_id": 858, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap17d053f9-c9", "ovs_interfaceid": "17d053f9-c98c-4b3d-8a11-1c514de74d81", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 710.897124] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Releasing lock "refresh_cache-2a329146-5e0d-4a1f-ba65-3efda2f5e1e0" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 710.897372] env[63531]: DEBUG nova.compute.manager [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 710.897522] env[63531]: DEBUG nova.compute.manager [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 710.897717] env[63531]: DEBUG nova.network.neutron [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 710.900429] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd133d04-11e3-4c8e-8657-6dd35a9a518b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.908141] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-519ab465-28d6-4791-b9a0-ad3735f6f0f9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.939943] env[63531]: DEBUG nova.network.neutron [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 710.941607] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54119f47-0469-40c5-a9d8-b203b2e0882c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.949484] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f497e583-555b-41a5-84e0-fc40ef423170 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.962831] env[63531]: DEBUG nova.compute.provider_tree [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 710.997845] env[63531]: DEBUG oslo_vmware.api [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5257dfc3-7b81-a5ad-05f1-53a58964b37c, 'name': SearchDatastore_Task, 'duration_secs': 0.009629} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.998115] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 710.998377] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] ffb34f90-d22f-440b-ba07-75d474c3c300/ffb34f90-d22f-440b-ba07-75d474c3c300.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 710.998665] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 710.998850] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 710.999070] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b457dfa1-74d1-4e05-869c-5a88d4c90a81 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.000958] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-838a1dbe-1bba-4b13-aa46-25ffa308ac4e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.009090] env[63531]: DEBUG oslo_vmware.api [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Waiting for the task: (returnval){ [ 711.009090] env[63531]: value = "task-1117535" [ 711.009090] env[63531]: _type = "Task" [ 711.009090] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.012188] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 711.012365] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 711.013318] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5c0a57aa-fc61-4375-9b08-2484a775dbc9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.018505] env[63531]: DEBUG oslo_vmware.api [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Task: {'id': task-1117535, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.021228] env[63531]: DEBUG oslo_vmware.api [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Waiting for the task: (returnval){ [ 711.021228] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]526131e7-023d-e210-8d42-35da8a96e13e" [ 711.021228] env[63531]: _type = "Task" [ 711.021228] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.021657] env[63531]: DEBUG oslo_concurrency.lockutils [req-b6e3d250-cd92-48fb-b457-4253ce852a14 req-e61f762d-af4c-4452-a7b8-59995bb835c3 service nova] Releasing lock "refresh_cache-ffb34f90-d22f-440b-ba07-75d474c3c300" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 711.028947] env[63531]: DEBUG oslo_vmware.api [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]526131e7-023d-e210-8d42-35da8a96e13e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.445401] env[63531]: DEBUG nova.network.neutron [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 711.465825] env[63531]: DEBUG nova.scheduler.client.report [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 711.518334] env[63531]: DEBUG oslo_vmware.api [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Task: {'id': task-1117535, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.503431} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.518553] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] ffb34f90-d22f-440b-ba07-75d474c3c300/ffb34f90-d22f-440b-ba07-75d474c3c300.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 711.518772] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: ffb34f90-d22f-440b-ba07-75d474c3c300] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 711.519041] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bb1006fd-cf84-4fbe-b516-9bd67c10b4a3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.526165] env[63531]: DEBUG oslo_vmware.api [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Waiting for the task: (returnval){ [ 711.526165] env[63531]: value = "task-1117536" [ 711.526165] env[63531]: _type = "Task" [ 711.526165] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.532537] env[63531]: DEBUG oslo_vmware.api [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]526131e7-023d-e210-8d42-35da8a96e13e, 'name': SearchDatastore_Task, 'duration_secs': 0.010961} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.533621] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-727c4f68-efed-456e-9bce-dad41d915146 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.538383] env[63531]: DEBUG oslo_vmware.api [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Task: {'id': task-1117536, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.541156] env[63531]: DEBUG oslo_vmware.api [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Waiting for the task: (returnval){ [ 711.541156] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5215a491-7c57-3f7d-6765-3d23f332dd1d" [ 711.541156] env[63531]: _type = "Task" [ 711.541156] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.548568] env[63531]: DEBUG oslo_vmware.api [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5215a491-7c57-3f7d-6765-3d23f332dd1d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.951054] env[63531]: INFO nova.compute.manager [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0] Took 1.05 seconds to deallocate network for instance. [ 711.970842] env[63531]: DEBUG oslo_concurrency.lockutils [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.367s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 711.971400] env[63531]: DEBUG nova.compute.manager [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] [instance: 7bc63fdd-4b50-486d-9681-ca0baa08f7b3] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 711.974753] env[63531]: DEBUG oslo_concurrency.lockutils [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.721s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 711.976196] env[63531]: INFO nova.compute.claims [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] [instance: f0276e34-1de3-482a-82c7-1439d40fd85e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 712.037313] env[63531]: DEBUG oslo_vmware.api [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Task: {'id': task-1117536, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.327966} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.037582] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: ffb34f90-d22f-440b-ba07-75d474c3c300] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 712.038363] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39436e29-3033-4d47-a5cd-9c25d9b2260a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.061861] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: ffb34f90-d22f-440b-ba07-75d474c3c300] Reconfiguring VM instance instance-00000020 to attach disk [datastore1] ffb34f90-d22f-440b-ba07-75d474c3c300/ffb34f90-d22f-440b-ba07-75d474c3c300.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 712.062607] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a2ca1e97-3c0c-4aa3-9ea2-fa93d8192805 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.079854] env[63531]: DEBUG oslo_vmware.api [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5215a491-7c57-3f7d-6765-3d23f332dd1d, 'name': SearchDatastore_Task, 'duration_secs': 0.009472} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.080447] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 712.080734] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 0c948655-362d-4673-b3e6-3f28ec69ea3c/0c948655-362d-4673-b3e6-3f28ec69ea3c.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 712.080981] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f4aac973-5724-40eb-96e1-09390634e8e4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.084909] env[63531]: DEBUG oslo_vmware.api [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Waiting for the task: (returnval){ [ 712.084909] env[63531]: value = "task-1117537" [ 712.084909] env[63531]: _type = "Task" [ 712.084909] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.089224] env[63531]: DEBUG oslo_vmware.api [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Waiting for the task: (returnval){ [ 712.089224] env[63531]: value = "task-1117538" [ 712.089224] env[63531]: _type = "Task" [ 712.089224] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.095217] env[63531]: DEBUG oslo_vmware.api [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Task: {'id': task-1117537, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.100943] env[63531]: DEBUG oslo_vmware.api [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Task: {'id': task-1117538, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.476523] env[63531]: DEBUG nova.compute.utils [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 712.477650] env[63531]: DEBUG nova.compute.manager [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] [instance: 7bc63fdd-4b50-486d-9681-ca0baa08f7b3] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 712.477865] env[63531]: DEBUG nova.network.neutron [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] [instance: 7bc63fdd-4b50-486d-9681-ca0baa08f7b3] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 712.523819] env[63531]: DEBUG nova.policy [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6f9beab14b754f0989096a31e3c61eaf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '974bca3f226949c29a9be3624ac5a299', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 712.596436] env[63531]: DEBUG oslo_vmware.api [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Task: {'id': task-1117537, 'name': ReconfigVM_Task, 'duration_secs': 0.342721} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.597137] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: ffb34f90-d22f-440b-ba07-75d474c3c300] Reconfigured VM instance instance-00000020 to attach disk [datastore1] ffb34f90-d22f-440b-ba07-75d474c3c300/ffb34f90-d22f-440b-ba07-75d474c3c300.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 712.597780] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6091f155-0de1-4c1e-8823-d8c592d72377 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.602133] env[63531]: DEBUG oslo_vmware.api [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Task: {'id': task-1117538, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.498594} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.602654] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 0c948655-362d-4673-b3e6-3f28ec69ea3c/0c948655-362d-4673-b3e6-3f28ec69ea3c.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 712.602856] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 712.603088] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5486b20c-10d9-4442-a031-583194745dbd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.606046] env[63531]: DEBUG oslo_vmware.api [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Waiting for the task: (returnval){ [ 712.606046] env[63531]: value = "task-1117539" [ 712.606046] env[63531]: _type = "Task" [ 712.606046] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.610073] env[63531]: DEBUG oslo_vmware.api [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Waiting for the task: (returnval){ [ 712.610073] env[63531]: value = "task-1117540" [ 712.610073] env[63531]: _type = "Task" [ 712.610073] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.616400] env[63531]: DEBUG oslo_vmware.api [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Task: {'id': task-1117539, 'name': Rename_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.621453] env[63531]: DEBUG oslo_vmware.api [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Task: {'id': task-1117540, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.777914] env[63531]: DEBUG nova.network.neutron [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] [instance: 7bc63fdd-4b50-486d-9681-ca0baa08f7b3] Successfully created port: 7348aaf4-66f3-47a2-9467-83aa99ae9fe4 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 712.983758] env[63531]: DEBUG nova.compute.manager [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] [instance: 7bc63fdd-4b50-486d-9681-ca0baa08f7b3] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 712.995068] env[63531]: INFO nova.scheduler.client.report [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Deleted allocations for instance 2a329146-5e0d-4a1f-ba65-3efda2f5e1e0 [ 713.121934] env[63531]: DEBUG oslo_vmware.api [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Task: {'id': task-1117540, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.107577} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.125620] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 713.125970] env[63531]: DEBUG oslo_vmware.api [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Task: {'id': task-1117539, 'name': Rename_Task, 'duration_secs': 0.183114} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.126800] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a891045e-7aab-43bc-94bc-b84cd873fca3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.129950] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: ffb34f90-d22f-440b-ba07-75d474c3c300] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 713.131835] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-17656b05-7da6-4c26-9f9c-8009a44c281b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.160192] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Reconfiguring VM instance instance-00000021 to attach disk [datastore1] 0c948655-362d-4673-b3e6-3f28ec69ea3c/0c948655-362d-4673-b3e6-3f28ec69ea3c.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 713.166112] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b33771cb-e83e-48ef-b564-c81613690d5a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.185026] env[63531]: DEBUG oslo_vmware.api [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Waiting for the task: (returnval){ [ 713.185026] env[63531]: value = "task-1117541" [ 713.185026] env[63531]: _type = "Task" [ 713.185026] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.188248] env[63531]: DEBUG oslo_vmware.api [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Waiting for the task: (returnval){ [ 713.188248] env[63531]: value = "task-1117542" [ 713.188248] env[63531]: _type = "Task" [ 713.188248] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.196975] env[63531]: DEBUG oslo_vmware.api [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Task: {'id': task-1117541, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.202161] env[63531]: DEBUG oslo_vmware.api [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Task: {'id': task-1117542, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.383654] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0e39bd9-7462-4b53-8783-1b0b5a117824 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.391425] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-454a986b-e2ec-4eee-a0de-a1f10b082fbf {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.426138] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2e8c362-3750-4c12-b2ff-85124c008697 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.434563] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f14f1859-68bc-4e46-a746-9f24dd16014b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.449684] env[63531]: DEBUG nova.compute.provider_tree [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 713.506142] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b57b2f86-d8e9-4d15-a1a2-9bda5c471047 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Lock "2a329146-5e0d-4a1f-ba65-3efda2f5e1e0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 118.247s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 713.696900] env[63531]: DEBUG oslo_vmware.api [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Task: {'id': task-1117541, 'name': PowerOnVM_Task, 'duration_secs': 0.507685} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.699921] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: ffb34f90-d22f-440b-ba07-75d474c3c300] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 713.700142] env[63531]: INFO nova.compute.manager [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: ffb34f90-d22f-440b-ba07-75d474c3c300] Took 6.39 seconds to spawn the instance on the hypervisor. [ 713.700323] env[63531]: DEBUG nova.compute.manager [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: ffb34f90-d22f-440b-ba07-75d474c3c300] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 713.700593] env[63531]: DEBUG oslo_vmware.api [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Task: {'id': task-1117542, 'name': ReconfigVM_Task, 'duration_secs': 0.301239} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.701278] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bb47791-5fe3-4607-9c78-0d09cff18ee1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.703668] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Reconfigured VM instance instance-00000021 to attach disk [datastore1] 0c948655-362d-4673-b3e6-3f28ec69ea3c/0c948655-362d-4673-b3e6-3f28ec69ea3c.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 713.704237] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-873a46b2-87e2-4cd1-9b22-ae6716a7ca6f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.713274] env[63531]: DEBUG oslo_vmware.api [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Waiting for the task: (returnval){ [ 713.713274] env[63531]: value = "task-1117543" [ 713.713274] env[63531]: _type = "Task" [ 713.713274] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.721528] env[63531]: DEBUG oslo_vmware.api [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Task: {'id': task-1117543, 'name': Rename_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.953176] env[63531]: DEBUG nova.scheduler.client.report [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 713.998240] env[63531]: DEBUG nova.compute.manager [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] [instance: 7bc63fdd-4b50-486d-9681-ca0baa08f7b3] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 714.008329] env[63531]: DEBUG nova.compute.manager [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 714.023685] env[63531]: DEBUG nova.virt.hardware [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 714.023931] env[63531]: DEBUG nova.virt.hardware [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 714.024104] env[63531]: DEBUG nova.virt.hardware [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 714.024295] env[63531]: DEBUG nova.virt.hardware [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 714.024443] env[63531]: DEBUG nova.virt.hardware [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 714.024587] env[63531]: DEBUG nova.virt.hardware [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 714.024793] env[63531]: DEBUG nova.virt.hardware [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 714.024948] env[63531]: DEBUG nova.virt.hardware [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 714.025130] env[63531]: DEBUG nova.virt.hardware [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 714.025285] env[63531]: DEBUG nova.virt.hardware [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 714.025452] env[63531]: DEBUG nova.virt.hardware [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 714.026318] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32b999b5-044a-46f3-bd7f-8fb446aa3af6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.038017] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-896325f7-1486-49fc-bdb1-8fbfaa015e4e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.231119] env[63531]: INFO nova.compute.manager [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: ffb34f90-d22f-440b-ba07-75d474c3c300] Took 40.47 seconds to build instance. [ 714.242239] env[63531]: DEBUG oslo_vmware.api [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Task: {'id': task-1117543, 'name': Rename_Task, 'duration_secs': 0.146989} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.243963] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 714.243963] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ff554981-2da3-4d4c-8898-9b704f184531 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.254017] env[63531]: DEBUG oslo_vmware.api [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Waiting for the task: (returnval){ [ 714.254017] env[63531]: value = "task-1117544" [ 714.254017] env[63531]: _type = "Task" [ 714.254017] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.259689] env[63531]: DEBUG nova.compute.manager [req-a5801f06-6728-42b5-b3e7-10d08e55ca05 req-c037ad05-d91f-4fe3-815d-08aa1890d4e9 service nova] [instance: 7bc63fdd-4b50-486d-9681-ca0baa08f7b3] Received event network-vif-plugged-7348aaf4-66f3-47a2-9467-83aa99ae9fe4 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 714.260084] env[63531]: DEBUG oslo_concurrency.lockutils [req-a5801f06-6728-42b5-b3e7-10d08e55ca05 req-c037ad05-d91f-4fe3-815d-08aa1890d4e9 service nova] Acquiring lock "7bc63fdd-4b50-486d-9681-ca0baa08f7b3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 714.260437] env[63531]: DEBUG oslo_concurrency.lockutils [req-a5801f06-6728-42b5-b3e7-10d08e55ca05 req-c037ad05-d91f-4fe3-815d-08aa1890d4e9 service nova] Lock "7bc63fdd-4b50-486d-9681-ca0baa08f7b3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 714.260851] env[63531]: DEBUG oslo_concurrency.lockutils [req-a5801f06-6728-42b5-b3e7-10d08e55ca05 req-c037ad05-d91f-4fe3-815d-08aa1890d4e9 service nova] Lock "7bc63fdd-4b50-486d-9681-ca0baa08f7b3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 714.261733] env[63531]: DEBUG nova.compute.manager [req-a5801f06-6728-42b5-b3e7-10d08e55ca05 req-c037ad05-d91f-4fe3-815d-08aa1890d4e9 service nova] [instance: 7bc63fdd-4b50-486d-9681-ca0baa08f7b3] No waiting events found dispatching network-vif-plugged-7348aaf4-66f3-47a2-9467-83aa99ae9fe4 {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 714.261733] env[63531]: WARNING nova.compute.manager [req-a5801f06-6728-42b5-b3e7-10d08e55ca05 req-c037ad05-d91f-4fe3-815d-08aa1890d4e9 service nova] [instance: 7bc63fdd-4b50-486d-9681-ca0baa08f7b3] Received unexpected event network-vif-plugged-7348aaf4-66f3-47a2-9467-83aa99ae9fe4 for instance with vm_state building and task_state spawning. [ 714.266892] env[63531]: DEBUG oslo_vmware.api [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Task: {'id': task-1117544, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.307362] env[63531]: DEBUG nova.network.neutron [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] [instance: 7bc63fdd-4b50-486d-9681-ca0baa08f7b3] Successfully updated port: 7348aaf4-66f3-47a2-9467-83aa99ae9fe4 {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 714.458959] env[63531]: DEBUG oslo_concurrency.lockutils [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.484s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 714.459511] env[63531]: DEBUG nova.compute.manager [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] [instance: f0276e34-1de3-482a-82c7-1439d40fd85e] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 714.462480] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.194s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 714.463839] env[63531]: INFO nova.compute.claims [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 714.536916] env[63531]: DEBUG oslo_concurrency.lockutils [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 714.743632] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4d76ef3d-553c-4261-a946-ca69ba970b25 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Lock "ffb34f90-d22f-440b-ba07-75d474c3c300" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 116.483s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 714.762361] env[63531]: DEBUG oslo_vmware.api [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Task: {'id': task-1117544, 'name': PowerOnVM_Task, 'duration_secs': 0.433684} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.762629] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 714.762838] env[63531]: INFO nova.compute.manager [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Took 5.05 seconds to spawn the instance on the hypervisor. [ 714.762993] env[63531]: DEBUG nova.compute.manager [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 714.764041] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ba829ad-5346-4b43-9db9-7041ff26feeb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.809837] env[63531]: DEBUG oslo_concurrency.lockutils [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Acquiring lock "refresh_cache-7bc63fdd-4b50-486d-9681-ca0baa08f7b3" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 714.810015] env[63531]: DEBUG oslo_concurrency.lockutils [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Acquired lock "refresh_cache-7bc63fdd-4b50-486d-9681-ca0baa08f7b3" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 714.810165] env[63531]: DEBUG nova.network.neutron [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] [instance: 7bc63fdd-4b50-486d-9681-ca0baa08f7b3] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 714.968688] env[63531]: DEBUG nova.compute.utils [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 714.975019] env[63531]: DEBUG nova.compute.manager [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] [instance: f0276e34-1de3-482a-82c7-1439d40fd85e] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 714.975019] env[63531]: DEBUG nova.network.neutron [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] [instance: f0276e34-1de3-482a-82c7-1439d40fd85e] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 715.017240] env[63531]: DEBUG nova.policy [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b25397c86fc423286d869ad34c7b9d8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7c20b4e45b9c444db3b051e03ce5181c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 715.246271] env[63531]: DEBUG nova.compute.manager [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 715.277941] env[63531]: DEBUG nova.network.neutron [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] [instance: f0276e34-1de3-482a-82c7-1439d40fd85e] Successfully created port: 08a2f25f-31d7-4af6-9116-f946e065155a {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 715.282996] env[63531]: INFO nova.compute.manager [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Took 39.58 seconds to build instance. [ 715.353642] env[63531]: DEBUG nova.network.neutron [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] [instance: 7bc63fdd-4b50-486d-9681-ca0baa08f7b3] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 715.473124] env[63531]: DEBUG nova.compute.manager [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] [instance: f0276e34-1de3-482a-82c7-1439d40fd85e] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 715.576436] env[63531]: DEBUG nova.network.neutron [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] [instance: 7bc63fdd-4b50-486d-9681-ca0baa08f7b3] Updating instance_info_cache with network_info: [{"id": "7348aaf4-66f3-47a2-9467-83aa99ae9fe4", "address": "fa:16:3e:fe:29:6a", "network": {"id": "7f21bb8c-d602-4ef8-b385-f80f4ca8e642", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-343094373-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "974bca3f226949c29a9be3624ac5a299", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0ed262a3-f84f-4e1c-bbaf-c6a10e0c243e", "external-id": "nsx-vlan-transportzone-146", "segmentation_id": 146, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7348aaf4-66", "ovs_interfaceid": "7348aaf4-66f3-47a2-9467-83aa99ae9fe4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.768576] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 715.785484] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e44d0bef-3e6c-4e08-b88e-a703fdc4893a tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Lock "0c948655-362d-4673-b3e6-3f28ec69ea3c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 116.055s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 715.830567] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8aac6d23-0be9-403d-a6e5-8221c8d42a8b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.838755] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3314616d-2342-4d12-aa1a-5d95ddf103ef {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.871742] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-675a8fba-c17d-447b-ac62-bb80a3b450e8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.880449] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5bd5cf0-6434-4798-9fe4-4ffa467d8470 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.894242] env[63531]: DEBUG nova.compute.provider_tree [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 715.901135] env[63531]: INFO nova.compute.manager [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Rebuilding instance [ 715.946026] env[63531]: DEBUG nova.compute.manager [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 715.946912] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-569086d1-93e1-4562-adb6-fb5a77470218 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.078658] env[63531]: DEBUG oslo_concurrency.lockutils [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Releasing lock "refresh_cache-7bc63fdd-4b50-486d-9681-ca0baa08f7b3" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 716.079072] env[63531]: DEBUG nova.compute.manager [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] [instance: 7bc63fdd-4b50-486d-9681-ca0baa08f7b3] Instance network_info: |[{"id": "7348aaf4-66f3-47a2-9467-83aa99ae9fe4", "address": "fa:16:3e:fe:29:6a", "network": {"id": "7f21bb8c-d602-4ef8-b385-f80f4ca8e642", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-343094373-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "974bca3f226949c29a9be3624ac5a299", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0ed262a3-f84f-4e1c-bbaf-c6a10e0c243e", "external-id": "nsx-vlan-transportzone-146", "segmentation_id": 146, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7348aaf4-66", "ovs_interfaceid": "7348aaf4-66f3-47a2-9467-83aa99ae9fe4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 716.079596] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] [instance: 7bc63fdd-4b50-486d-9681-ca0baa08f7b3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fe:29:6a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0ed262a3-f84f-4e1c-bbaf-c6a10e0c243e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7348aaf4-66f3-47a2-9467-83aa99ae9fe4', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 716.087515] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Creating folder: Project (974bca3f226949c29a9be3624ac5a299). Parent ref: group-v244585. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 716.088118] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4a376e84-8a18-462e-82de-ab244465d913 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.094013] env[63531]: DEBUG oslo_concurrency.lockutils [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Acquiring lock "2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 716.094700] env[63531]: DEBUG oslo_concurrency.lockutils [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Lock "2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 716.099659] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Created folder: Project (974bca3f226949c29a9be3624ac5a299) in parent group-v244585. [ 716.099886] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Creating folder: Instances. Parent ref: group-v244608. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 716.100156] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-971b6d11-9644-497c-b6f5-cb1a515bcd58 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.109651] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Created folder: Instances in parent group-v244608. [ 716.109943] env[63531]: DEBUG oslo.service.loopingcall [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 716.110188] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7bc63fdd-4b50-486d-9681-ca0baa08f7b3] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 716.111217] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ab725ef6-dc5f-4e3e-8324-8e442f9b60e2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.130207] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 716.130207] env[63531]: value = "task-1117547" [ 716.130207] env[63531]: _type = "Task" [ 716.130207] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.138118] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117547, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.288058] env[63531]: DEBUG nova.compute.manager [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 716.293091] env[63531]: DEBUG nova.compute.manager [req-d20c6e5d-19f9-4469-ada7-c97d9bd05fca req-49dc5de8-1392-4430-b58e-ab741d55b1d5 service nova] [instance: 7bc63fdd-4b50-486d-9681-ca0baa08f7b3] Received event network-changed-7348aaf4-66f3-47a2-9467-83aa99ae9fe4 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 716.293305] env[63531]: DEBUG nova.compute.manager [req-d20c6e5d-19f9-4469-ada7-c97d9bd05fca req-49dc5de8-1392-4430-b58e-ab741d55b1d5 service nova] [instance: 7bc63fdd-4b50-486d-9681-ca0baa08f7b3] Refreshing instance network info cache due to event network-changed-7348aaf4-66f3-47a2-9467-83aa99ae9fe4. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 716.293539] env[63531]: DEBUG oslo_concurrency.lockutils [req-d20c6e5d-19f9-4469-ada7-c97d9bd05fca req-49dc5de8-1392-4430-b58e-ab741d55b1d5 service nova] Acquiring lock "refresh_cache-7bc63fdd-4b50-486d-9681-ca0baa08f7b3" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 716.293704] env[63531]: DEBUG oslo_concurrency.lockutils [req-d20c6e5d-19f9-4469-ada7-c97d9bd05fca req-49dc5de8-1392-4430-b58e-ab741d55b1d5 service nova] Acquired lock "refresh_cache-7bc63fdd-4b50-486d-9681-ca0baa08f7b3" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.293878] env[63531]: DEBUG nova.network.neutron [req-d20c6e5d-19f9-4469-ada7-c97d9bd05fca req-49dc5de8-1392-4430-b58e-ab741d55b1d5 service nova] [instance: 7bc63fdd-4b50-486d-9681-ca0baa08f7b3] Refreshing network info cache for port 7348aaf4-66f3-47a2-9467-83aa99ae9fe4 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 716.397197] env[63531]: DEBUG nova.scheduler.client.report [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 716.457607] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 716.458770] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bb841b5c-fea9-49b6-83d5-c031bc2b17ba {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.464544] env[63531]: DEBUG oslo_vmware.api [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Waiting for the task: (returnval){ [ 716.464544] env[63531]: value = "task-1117548" [ 716.464544] env[63531]: _type = "Task" [ 716.464544] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.474239] env[63531]: DEBUG oslo_vmware.api [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Task: {'id': task-1117548, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.482273] env[63531]: DEBUG nova.compute.manager [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] [instance: f0276e34-1de3-482a-82c7-1439d40fd85e] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 716.505967] env[63531]: DEBUG nova.virt.hardware [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 716.506225] env[63531]: DEBUG nova.virt.hardware [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 716.506379] env[63531]: DEBUG nova.virt.hardware [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 716.506658] env[63531]: DEBUG nova.virt.hardware [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 716.506838] env[63531]: DEBUG nova.virt.hardware [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 716.506988] env[63531]: DEBUG nova.virt.hardware [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 716.507556] env[63531]: DEBUG nova.virt.hardware [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 716.507556] env[63531]: DEBUG nova.virt.hardware [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 716.507556] env[63531]: DEBUG nova.virt.hardware [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 716.507751] env[63531]: DEBUG nova.virt.hardware [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 716.507824] env[63531]: DEBUG nova.virt.hardware [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 716.508674] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c67f10a-a166-4228-abd2-06c07fd805b1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.516151] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72e9f5f5-340c-47eb-ab3c-fa4866fe9079 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.639876] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117547, 'name': CreateVM_Task, 'duration_secs': 0.371224} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.640205] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7bc63fdd-4b50-486d-9681-ca0baa08f7b3] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 716.640729] env[63531]: DEBUG oslo_concurrency.lockutils [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 716.640898] env[63531]: DEBUG oslo_concurrency.lockutils [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.641256] env[63531]: DEBUG oslo_concurrency.lockutils [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 716.641483] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b339cd0-38fd-4eb8-b1d5-ce6ee6ca2504 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.646304] env[63531]: DEBUG oslo_vmware.api [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Waiting for the task: (returnval){ [ 716.646304] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52c2de00-3a75-2b33-2cba-c9a5d6fe1357" [ 716.646304] env[63531]: _type = "Task" [ 716.646304] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.654272] env[63531]: DEBUG oslo_vmware.api [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52c2de00-3a75-2b33-2cba-c9a5d6fe1357, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.819382] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 716.902465] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.440s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 716.903016] env[63531]: DEBUG nova.compute.manager [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 716.908385] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.662s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 716.974746] env[63531]: DEBUG oslo_vmware.api [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Task: {'id': task-1117548, 'name': PowerOffVM_Task, 'duration_secs': 0.253495} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.975032] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 716.975263] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 716.976028] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fbfb3b6-0fbc-4901-b1a4-9f92a7d5afdc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.982972] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 716.983229] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-aaab82ae-7720-443f-98d4-064e124cdb76 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.007788] env[63531]: DEBUG nova.network.neutron [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] [instance: f0276e34-1de3-482a-82c7-1439d40fd85e] Successfully updated port: 08a2f25f-31d7-4af6-9116-f946e065155a {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 717.011928] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 717.012312] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 717.012467] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Deleting the datastore file [datastore1] 0c948655-362d-4673-b3e6-3f28ec69ea3c {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 717.013493] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a46abe4c-a871-45d8-9734-80e98a1b64bc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.020460] env[63531]: DEBUG oslo_vmware.api [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Waiting for the task: (returnval){ [ 717.020460] env[63531]: value = "task-1117550" [ 717.020460] env[63531]: _type = "Task" [ 717.020460] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.033987] env[63531]: DEBUG oslo_vmware.api [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Task: {'id': task-1117550, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.104643] env[63531]: DEBUG nova.network.neutron [req-d20c6e5d-19f9-4469-ada7-c97d9bd05fca req-49dc5de8-1392-4430-b58e-ab741d55b1d5 service nova] [instance: 7bc63fdd-4b50-486d-9681-ca0baa08f7b3] Updated VIF entry in instance network info cache for port 7348aaf4-66f3-47a2-9467-83aa99ae9fe4. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 717.105349] env[63531]: DEBUG nova.network.neutron [req-d20c6e5d-19f9-4469-ada7-c97d9bd05fca req-49dc5de8-1392-4430-b58e-ab741d55b1d5 service nova] [instance: 7bc63fdd-4b50-486d-9681-ca0baa08f7b3] Updating instance_info_cache with network_info: [{"id": "7348aaf4-66f3-47a2-9467-83aa99ae9fe4", "address": "fa:16:3e:fe:29:6a", "network": {"id": "7f21bb8c-d602-4ef8-b385-f80f4ca8e642", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-343094373-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "974bca3f226949c29a9be3624ac5a299", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0ed262a3-f84f-4e1c-bbaf-c6a10e0c243e", "external-id": "nsx-vlan-transportzone-146", "segmentation_id": 146, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7348aaf4-66", "ovs_interfaceid": "7348aaf4-66f3-47a2-9467-83aa99ae9fe4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 717.157031] env[63531]: DEBUG oslo_vmware.api [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52c2de00-3a75-2b33-2cba-c9a5d6fe1357, 'name': SearchDatastore_Task, 'duration_secs': 0.009833} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.157389] env[63531]: DEBUG oslo_concurrency.lockutils [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 717.157668] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] [instance: 7bc63fdd-4b50-486d-9681-ca0baa08f7b3] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 717.157949] env[63531]: DEBUG oslo_concurrency.lockutils [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 717.158149] env[63531]: DEBUG oslo_concurrency.lockutils [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.158369] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 717.158655] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8110fbe4-e086-4335-acf1-f35415a1cf16 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.166470] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 717.167576] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 717.167576] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f3c986c6-c89c-418e-8e40-aca9f720a968 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.172386] env[63531]: DEBUG oslo_vmware.api [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Waiting for the task: (returnval){ [ 717.172386] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52ba2884-b73f-29af-3453-40a0a9d89c29" [ 717.172386] env[63531]: _type = "Task" [ 717.172386] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.179802] env[63531]: DEBUG oslo_vmware.api [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52ba2884-b73f-29af-3453-40a0a9d89c29, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.413608] env[63531]: DEBUG nova.compute.utils [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 717.417997] env[63531]: DEBUG nova.compute.manager [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 717.418176] env[63531]: DEBUG nova.network.neutron [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 717.456796] env[63531]: DEBUG nova.policy [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '43a22fda8420440eb0e60eab86eaed1a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e2efa9eb41854d9cbcf96ebb8edbcfc9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 717.512786] env[63531]: DEBUG oslo_concurrency.lockutils [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Acquiring lock "refresh_cache-f0276e34-1de3-482a-82c7-1439d40fd85e" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 717.512978] env[63531]: DEBUG oslo_concurrency.lockutils [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Acquired lock "refresh_cache-f0276e34-1de3-482a-82c7-1439d40fd85e" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.513281] env[63531]: DEBUG nova.network.neutron [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] [instance: f0276e34-1de3-482a-82c7-1439d40fd85e] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 717.532839] env[63531]: DEBUG oslo_vmware.api [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Task: {'id': task-1117550, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.106358} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.534077] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 717.534077] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 717.534077] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 717.609970] env[63531]: DEBUG oslo_concurrency.lockutils [req-d20c6e5d-19f9-4469-ada7-c97d9bd05fca req-49dc5de8-1392-4430-b58e-ab741d55b1d5 service nova] Releasing lock "refresh_cache-7bc63fdd-4b50-486d-9681-ca0baa08f7b3" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 717.684751] env[63531]: DEBUG oslo_vmware.api [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52ba2884-b73f-29af-3453-40a0a9d89c29, 'name': SearchDatastore_Task, 'duration_secs': 0.007988} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.686089] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-05f44668-4f19-46fb-b1ea-e91ab0bebc55 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.691204] env[63531]: DEBUG oslo_vmware.api [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Waiting for the task: (returnval){ [ 717.691204] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]528e80a7-c76e-e5f3-8bd9-7ccc165a0a85" [ 717.691204] env[63531]: _type = "Task" [ 717.691204] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.700895] env[63531]: DEBUG oslo_vmware.api [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]528e80a7-c76e-e5f3-8bd9-7ccc165a0a85, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.725875] env[63531]: DEBUG nova.network.neutron [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Successfully created port: 0c5ad75f-1b78-47fd-bd2e-331a7ae0732a {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 717.769635] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec3e261a-e226-41dc-ade9-8044ffa40a80 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.777895] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87d0f4e1-7a40-4762-8561-605ef048ec3b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.809634] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d54304d-b2ac-4ec2-9497-e6fa2f22b279 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.817123] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b818a1b-8d8d-4e0d-a0e7-2c721b35ce52 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.831380] env[63531]: DEBUG nova.compute.provider_tree [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 717.922747] env[63531]: DEBUG nova.compute.manager [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 718.045576] env[63531]: DEBUG nova.network.neutron [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] [instance: f0276e34-1de3-482a-82c7-1439d40fd85e] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 718.178606] env[63531]: DEBUG nova.network.neutron [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] [instance: f0276e34-1de3-482a-82c7-1439d40fd85e] Updating instance_info_cache with network_info: [{"id": "08a2f25f-31d7-4af6-9116-f946e065155a", "address": "fa:16:3e:6b:21:cd", "network": {"id": "2ac03a43-35b9-4b28-8e3b-b7c1435b57d6", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-1401500218-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7c20b4e45b9c444db3b051e03ce5181c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "32028d02-abaa-4071-bc65-1460f5c772a8", "external-id": "nsx-vlan-transportzone-558", "segmentation_id": 558, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap08a2f25f-31", "ovs_interfaceid": "08a2f25f-31d7-4af6-9116-f946e065155a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.201609] env[63531]: DEBUG oslo_vmware.api [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]528e80a7-c76e-e5f3-8bd9-7ccc165a0a85, 'name': SearchDatastore_Task, 'duration_secs': 0.008838} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.201855] env[63531]: DEBUG oslo_concurrency.lockutils [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 718.202153] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 7bc63fdd-4b50-486d-9681-ca0baa08f7b3/7bc63fdd-4b50-486d-9681-ca0baa08f7b3.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 718.202405] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-02a292c9-d914-47b4-b904-0b17d1bc9b12 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.210048] env[63531]: DEBUG oslo_vmware.api [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Waiting for the task: (returnval){ [ 718.210048] env[63531]: value = "task-1117551" [ 718.210048] env[63531]: _type = "Task" [ 718.210048] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.217459] env[63531]: DEBUG oslo_vmware.api [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Task: {'id': task-1117551, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.318427] env[63531]: DEBUG nova.compute.manager [req-7d5ddd3c-3602-4e04-8a8b-b785ed5c3634 req-22e8ddee-970f-4adb-bf66-031a17940234 service nova] [instance: f0276e34-1de3-482a-82c7-1439d40fd85e] Received event network-vif-plugged-08a2f25f-31d7-4af6-9116-f946e065155a {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 718.318714] env[63531]: DEBUG oslo_concurrency.lockutils [req-7d5ddd3c-3602-4e04-8a8b-b785ed5c3634 req-22e8ddee-970f-4adb-bf66-031a17940234 service nova] Acquiring lock "f0276e34-1de3-482a-82c7-1439d40fd85e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 718.318930] env[63531]: DEBUG oslo_concurrency.lockutils [req-7d5ddd3c-3602-4e04-8a8b-b785ed5c3634 req-22e8ddee-970f-4adb-bf66-031a17940234 service nova] Lock "f0276e34-1de3-482a-82c7-1439d40fd85e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 718.319121] env[63531]: DEBUG oslo_concurrency.lockutils [req-7d5ddd3c-3602-4e04-8a8b-b785ed5c3634 req-22e8ddee-970f-4adb-bf66-031a17940234 service nova] Lock "f0276e34-1de3-482a-82c7-1439d40fd85e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 718.319325] env[63531]: DEBUG nova.compute.manager [req-7d5ddd3c-3602-4e04-8a8b-b785ed5c3634 req-22e8ddee-970f-4adb-bf66-031a17940234 service nova] [instance: f0276e34-1de3-482a-82c7-1439d40fd85e] No waiting events found dispatching network-vif-plugged-08a2f25f-31d7-4af6-9116-f946e065155a {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 718.319525] env[63531]: WARNING nova.compute.manager [req-7d5ddd3c-3602-4e04-8a8b-b785ed5c3634 req-22e8ddee-970f-4adb-bf66-031a17940234 service nova] [instance: f0276e34-1de3-482a-82c7-1439d40fd85e] Received unexpected event network-vif-plugged-08a2f25f-31d7-4af6-9116-f946e065155a for instance with vm_state building and task_state spawning. [ 718.319721] env[63531]: DEBUG nova.compute.manager [req-7d5ddd3c-3602-4e04-8a8b-b785ed5c3634 req-22e8ddee-970f-4adb-bf66-031a17940234 service nova] [instance: f0276e34-1de3-482a-82c7-1439d40fd85e] Received event network-changed-08a2f25f-31d7-4af6-9116-f946e065155a {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 718.319915] env[63531]: DEBUG nova.compute.manager [req-7d5ddd3c-3602-4e04-8a8b-b785ed5c3634 req-22e8ddee-970f-4adb-bf66-031a17940234 service nova] [instance: f0276e34-1de3-482a-82c7-1439d40fd85e] Refreshing instance network info cache due to event network-changed-08a2f25f-31d7-4af6-9116-f946e065155a. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 718.320118] env[63531]: DEBUG oslo_concurrency.lockutils [req-7d5ddd3c-3602-4e04-8a8b-b785ed5c3634 req-22e8ddee-970f-4adb-bf66-031a17940234 service nova] Acquiring lock "refresh_cache-f0276e34-1de3-482a-82c7-1439d40fd85e" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 718.334764] env[63531]: DEBUG nova.scheduler.client.report [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 718.577999] env[63531]: DEBUG nova.virt.hardware [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 718.578412] env[63531]: DEBUG nova.virt.hardware [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 718.578673] env[63531]: DEBUG nova.virt.hardware [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 718.578955] env[63531]: DEBUG nova.virt.hardware [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 718.579215] env[63531]: DEBUG nova.virt.hardware [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 718.579491] env[63531]: DEBUG nova.virt.hardware [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 718.579807] env[63531]: DEBUG nova.virt.hardware [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 718.580079] env[63531]: DEBUG nova.virt.hardware [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 718.580321] env[63531]: DEBUG nova.virt.hardware [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 718.580631] env[63531]: DEBUG nova.virt.hardware [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 718.580889] env[63531]: DEBUG nova.virt.hardware [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 718.582214] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af27a2a4-1591-4ded-91cf-ea3a659f63d0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.593900] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cace10e-c676-4f8d-adbc-53f19bc6c7b4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.614442] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Instance VIF info [] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 718.622491] env[63531]: DEBUG oslo.service.loopingcall [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 718.622911] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 718.623252] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9c898fd5-c106-4fb6-9c6e-52ffdaf22f11 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.645767] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 718.645767] env[63531]: value = "task-1117552" [ 718.645767] env[63531]: _type = "Task" [ 718.645767] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.654057] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117552, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.680850] env[63531]: DEBUG oslo_concurrency.lockutils [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Releasing lock "refresh_cache-f0276e34-1de3-482a-82c7-1439d40fd85e" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 718.681203] env[63531]: DEBUG nova.compute.manager [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] [instance: f0276e34-1de3-482a-82c7-1439d40fd85e] Instance network_info: |[{"id": "08a2f25f-31d7-4af6-9116-f946e065155a", "address": "fa:16:3e:6b:21:cd", "network": {"id": "2ac03a43-35b9-4b28-8e3b-b7c1435b57d6", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-1401500218-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7c20b4e45b9c444db3b051e03ce5181c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "32028d02-abaa-4071-bc65-1460f5c772a8", "external-id": "nsx-vlan-transportzone-558", "segmentation_id": 558, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap08a2f25f-31", "ovs_interfaceid": "08a2f25f-31d7-4af6-9116-f946e065155a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 718.681499] env[63531]: DEBUG oslo_concurrency.lockutils [req-7d5ddd3c-3602-4e04-8a8b-b785ed5c3634 req-22e8ddee-970f-4adb-bf66-031a17940234 service nova] Acquired lock "refresh_cache-f0276e34-1de3-482a-82c7-1439d40fd85e" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.681676] env[63531]: DEBUG nova.network.neutron [req-7d5ddd3c-3602-4e04-8a8b-b785ed5c3634 req-22e8ddee-970f-4adb-bf66-031a17940234 service nova] [instance: f0276e34-1de3-482a-82c7-1439d40fd85e] Refreshing network info cache for port 08a2f25f-31d7-4af6-9116-f946e065155a {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 718.682936] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] [instance: f0276e34-1de3-482a-82c7-1439d40fd85e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6b:21:cd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '32028d02-abaa-4071-bc65-1460f5c772a8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '08a2f25f-31d7-4af6-9116-f946e065155a', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 718.691187] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Creating folder: Project (7c20b4e45b9c444db3b051e03ce5181c). Parent ref: group-v244585. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 718.692278] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9e9756d5-d5af-45d8-bb77-d88877e99890 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.702277] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Created folder: Project (7c20b4e45b9c444db3b051e03ce5181c) in parent group-v244585. [ 718.702492] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Creating folder: Instances. Parent ref: group-v244612. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 718.702728] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-16204d97-d0c4-4aff-b0a1-650ce6596d21 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.712140] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Created folder: Instances in parent group-v244612. [ 718.712140] env[63531]: DEBUG oslo.service.loopingcall [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 718.714993] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f0276e34-1de3-482a-82c7-1439d40fd85e] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 718.715223] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ecc771f0-2561-4170-b7af-7a48f2102c9e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.734416] env[63531]: DEBUG oslo_vmware.api [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Task: {'id': task-1117551, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.461691} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.735607] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 7bc63fdd-4b50-486d-9681-ca0baa08f7b3/7bc63fdd-4b50-486d-9681-ca0baa08f7b3.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 718.735838] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] [instance: 7bc63fdd-4b50-486d-9681-ca0baa08f7b3] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 718.736076] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 718.736076] env[63531]: value = "task-1117555" [ 718.736076] env[63531]: _type = "Task" [ 718.736076] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.736254] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-beadacf2-9651-4509-b6cf-998537edff94 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.746209] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117555, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.747394] env[63531]: DEBUG oslo_vmware.api [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Waiting for the task: (returnval){ [ 718.747394] env[63531]: value = "task-1117556" [ 718.747394] env[63531]: _type = "Task" [ 718.747394] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.756689] env[63531]: DEBUG oslo_vmware.api [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Task: {'id': task-1117556, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.840828] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.933s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 718.841479] env[63531]: ERROR nova.compute.manager [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: 3f3113dd-863f-442d-a214-28dde377d227] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d0cc21fa-d502-4f4b-99e7-ff4309c06ff5, please check neutron logs for more information. [ 718.841479] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] Traceback (most recent call last): [ 718.841479] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 718.841479] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] self.driver.spawn(context, instance, image_meta, [ 718.841479] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 718.841479] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] self._vmops.spawn(context, instance, image_meta, injected_files, [ 718.841479] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 718.841479] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] vm_ref = self.build_virtual_machine(instance, [ 718.841479] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 718.841479] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] vif_infos = vmwarevif.get_vif_info(self._session, [ 718.841479] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 718.841887] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] for vif in network_info: [ 718.841887] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 718.841887] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] return self._sync_wrapper(fn, *args, **kwargs) [ 718.841887] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 718.841887] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] self.wait() [ 718.841887] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 718.841887] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] self[:] = self._gt.wait() [ 718.841887] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 718.841887] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] return self._exit_event.wait() [ 718.841887] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 718.841887] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] current.throw(*self._exc) [ 718.841887] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 718.841887] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] result = function(*args, **kwargs) [ 718.842371] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 718.842371] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] return func(*args, **kwargs) [ 718.842371] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 718.842371] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] raise e [ 718.842371] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 718.842371] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] nwinfo = self.network_api.allocate_for_instance( [ 718.842371] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 718.842371] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] created_port_ids = self._update_ports_for_instance( [ 718.842371] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 718.842371] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] with excutils.save_and_reraise_exception(): [ 718.842371] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 718.842371] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] self.force_reraise() [ 718.842371] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 718.842777] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] raise self.value [ 718.842777] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 718.842777] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] updated_port = self._update_port( [ 718.842777] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 718.842777] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] _ensure_no_port_binding_failure(port) [ 718.842777] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 718.842777] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] raise exception.PortBindingFailed(port_id=port['id']) [ 718.842777] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] nova.exception.PortBindingFailed: Binding failed for port d0cc21fa-d502-4f4b-99e7-ff4309c06ff5, please check neutron logs for more information. [ 718.842777] env[63531]: ERROR nova.compute.manager [instance: 3f3113dd-863f-442d-a214-28dde377d227] [ 718.842777] env[63531]: DEBUG nova.compute.utils [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: 3f3113dd-863f-442d-a214-28dde377d227] Binding failed for port d0cc21fa-d502-4f4b-99e7-ff4309c06ff5, please check neutron logs for more information. {{(pid=63531) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 718.844037] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.059s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 718.845577] env[63531]: INFO nova.compute.claims [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: 645abb7a-cf73-4fb0-a9a0-49205f060d6f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 718.848558] env[63531]: DEBUG nova.compute.manager [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: 3f3113dd-863f-442d-a214-28dde377d227] Build of instance 3f3113dd-863f-442d-a214-28dde377d227 was re-scheduled: Binding failed for port d0cc21fa-d502-4f4b-99e7-ff4309c06ff5, please check neutron logs for more information. {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 718.849037] env[63531]: DEBUG nova.compute.manager [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: 3f3113dd-863f-442d-a214-28dde377d227] Unplugging VIFs for instance {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 718.849270] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Acquiring lock "refresh_cache-3f3113dd-863f-442d-a214-28dde377d227" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 718.849416] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Acquired lock "refresh_cache-3f3113dd-863f-442d-a214-28dde377d227" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.849571] env[63531]: DEBUG nova.network.neutron [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: 3f3113dd-863f-442d-a214-28dde377d227] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 718.933966] env[63531]: DEBUG nova.compute.manager [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 718.954706] env[63531]: DEBUG nova.virt.hardware [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 718.955537] env[63531]: DEBUG nova.virt.hardware [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 718.955537] env[63531]: DEBUG nova.virt.hardware [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 718.955537] env[63531]: DEBUG nova.virt.hardware [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 718.955537] env[63531]: DEBUG nova.virt.hardware [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 718.955741] env[63531]: DEBUG nova.virt.hardware [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 718.955844] env[63531]: DEBUG nova.virt.hardware [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 718.956011] env[63531]: DEBUG nova.virt.hardware [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 718.956189] env[63531]: DEBUG nova.virt.hardware [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 718.956350] env[63531]: DEBUG nova.virt.hardware [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 718.956526] env[63531]: DEBUG nova.virt.hardware [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 718.957738] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-462028ab-167a-4927-bcba-630daa20855a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.965669] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0dd1174-e02c-427a-a749-8e21dcb85f02 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.156522] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117552, 'name': CreateVM_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.246795] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117555, 'name': CreateVM_Task, 'duration_secs': 0.313932} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.246963] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f0276e34-1de3-482a-82c7-1439d40fd85e] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 719.247622] env[63531]: DEBUG oslo_concurrency.lockutils [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 719.247845] env[63531]: DEBUG oslo_concurrency.lockutils [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.248163] env[63531]: DEBUG oslo_concurrency.lockutils [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 719.248500] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d7283d1a-2b97-47b0-8700-9af07b68996b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.259271] env[63531]: DEBUG oslo_vmware.api [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Task: {'id': task-1117556, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.087616} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.259478] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] [instance: 7bc63fdd-4b50-486d-9681-ca0baa08f7b3] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 719.259849] env[63531]: DEBUG oslo_vmware.api [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Waiting for the task: (returnval){ [ 719.259849] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52045e63-6c0b-ffee-5339-9b76d4616f58" [ 719.259849] env[63531]: _type = "Task" [ 719.259849] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.260527] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-720cb5d2-041f-407f-9f87-90a84f631f60 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.285271] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] [instance: 7bc63fdd-4b50-486d-9681-ca0baa08f7b3] Reconfiguring VM instance instance-00000022 to attach disk [datastore1] 7bc63fdd-4b50-486d-9681-ca0baa08f7b3/7bc63fdd-4b50-486d-9681-ca0baa08f7b3.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 719.288816] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e2839062-2beb-4720-9b3a-cc7509f061f5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.310748] env[63531]: DEBUG oslo_vmware.api [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52045e63-6c0b-ffee-5339-9b76d4616f58, 'name': SearchDatastore_Task, 'duration_secs': 0.009369} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.313883] env[63531]: DEBUG oslo_concurrency.lockutils [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 719.314131] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] [instance: f0276e34-1de3-482a-82c7-1439d40fd85e] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 719.314349] env[63531]: DEBUG oslo_concurrency.lockutils [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 719.314546] env[63531]: DEBUG oslo_concurrency.lockutils [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.314709] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 719.315563] env[63531]: DEBUG nova.network.neutron [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Successfully updated port: 0c5ad75f-1b78-47fd-bd2e-331a7ae0732a {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 719.317255] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-578da166-e7c0-4ec8-ab30-39701a8fe37a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.322610] env[63531]: DEBUG oslo_vmware.api [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Waiting for the task: (returnval){ [ 719.322610] env[63531]: value = "task-1117557" [ 719.322610] env[63531]: _type = "Task" [ 719.322610] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.327609] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 719.327791] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 719.332228] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-efd0e218-48d7-431f-bc4e-64fb6315485d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.334694] env[63531]: DEBUG oslo_vmware.api [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Task: {'id': task-1117557, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.341615] env[63531]: DEBUG oslo_vmware.api [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Waiting for the task: (returnval){ [ 719.341615] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]522e5cb8-f151-1671-47e4-bc2bd00cce8e" [ 719.341615] env[63531]: _type = "Task" [ 719.341615] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.349847] env[63531]: DEBUG oslo_vmware.api [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]522e5cb8-f151-1671-47e4-bc2bd00cce8e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.373141] env[63531]: DEBUG nova.network.neutron [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: 3f3113dd-863f-442d-a214-28dde377d227] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 719.488156] env[63531]: DEBUG nova.network.neutron [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: 3f3113dd-863f-442d-a214-28dde377d227] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.508215] env[63531]: DEBUG nova.network.neutron [req-7d5ddd3c-3602-4e04-8a8b-b785ed5c3634 req-22e8ddee-970f-4adb-bf66-031a17940234 service nova] [instance: f0276e34-1de3-482a-82c7-1439d40fd85e] Updated VIF entry in instance network info cache for port 08a2f25f-31d7-4af6-9116-f946e065155a. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 719.508550] env[63531]: DEBUG nova.network.neutron [req-7d5ddd3c-3602-4e04-8a8b-b785ed5c3634 req-22e8ddee-970f-4adb-bf66-031a17940234 service nova] [instance: f0276e34-1de3-482a-82c7-1439d40fd85e] Updating instance_info_cache with network_info: [{"id": "08a2f25f-31d7-4af6-9116-f946e065155a", "address": "fa:16:3e:6b:21:cd", "network": {"id": "2ac03a43-35b9-4b28-8e3b-b7c1435b57d6", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-1401500218-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7c20b4e45b9c444db3b051e03ce5181c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "32028d02-abaa-4071-bc65-1460f5c772a8", "external-id": "nsx-vlan-transportzone-558", "segmentation_id": 558, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap08a2f25f-31", "ovs_interfaceid": "08a2f25f-31d7-4af6-9116-f946e065155a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.656871] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117552, 'name': CreateVM_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.822364] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Acquiring lock "refresh_cache-c3d597b4-1dc3-48d8-9bee-e73c8929181b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 719.822364] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Acquired lock "refresh_cache-c3d597b4-1dc3-48d8-9bee-e73c8929181b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.822364] env[63531]: DEBUG nova.network.neutron [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 719.833459] env[63531]: DEBUG oslo_vmware.api [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Task: {'id': task-1117557, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.850090] env[63531]: DEBUG oslo_vmware.api [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]522e5cb8-f151-1671-47e4-bc2bd00cce8e, 'name': SearchDatastore_Task, 'duration_secs': 0.009876} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.850898] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1c83e084-b03b-40b6-91bc-bb9dd1ae6569 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.858892] env[63531]: DEBUG oslo_vmware.api [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Waiting for the task: (returnval){ [ 719.858892] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]529dba0f-b92e-f327-08ad-13b9dc7ca13a" [ 719.858892] env[63531]: _type = "Task" [ 719.858892] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.867143] env[63531]: DEBUG oslo_vmware.api [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]529dba0f-b92e-f327-08ad-13b9dc7ca13a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.993393] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Releasing lock "refresh_cache-3f3113dd-863f-442d-a214-28dde377d227" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 719.993642] env[63531]: DEBUG nova.compute.manager [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63531) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 719.993830] env[63531]: DEBUG nova.compute.manager [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: 3f3113dd-863f-442d-a214-28dde377d227] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 719.994000] env[63531]: DEBUG nova.network.neutron [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: 3f3113dd-863f-442d-a214-28dde377d227] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 720.009659] env[63531]: DEBUG nova.network.neutron [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: 3f3113dd-863f-442d-a214-28dde377d227] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 720.011276] env[63531]: DEBUG oslo_concurrency.lockutils [req-7d5ddd3c-3602-4e04-8a8b-b785ed5c3634 req-22e8ddee-970f-4adb-bf66-031a17940234 service nova] Releasing lock "refresh_cache-f0276e34-1de3-482a-82c7-1439d40fd85e" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 720.158936] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117552, 'name': CreateVM_Task, 'duration_secs': 1.290828} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.159786] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 720.160522] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71a68ac2-6065-458f-8387-78d966822c22 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.163109] env[63531]: DEBUG oslo_concurrency.lockutils [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 720.163288] env[63531]: DEBUG oslo_concurrency.lockutils [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 720.165667] env[63531]: DEBUG oslo_concurrency.lockutils [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 720.165667] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4cfdd967-905e-4632-85e7-4368b1bac069 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.172043] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a2510a2-ae7a-4d0c-86e5-373a82237987 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.175125] env[63531]: DEBUG oslo_vmware.api [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Waiting for the task: (returnval){ [ 720.175125] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52faf1f1-9da7-f34f-1de2-e84d256e5362" [ 720.175125] env[63531]: _type = "Task" [ 720.175125] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.206537] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d84e486-58e9-4910-8a17-6f5d1007f5ab {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.213659] env[63531]: DEBUG oslo_vmware.api [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52faf1f1-9da7-f34f-1de2-e84d256e5362, 'name': SearchDatastore_Task, 'duration_secs': 0.008524} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.213925] env[63531]: DEBUG oslo_concurrency.lockutils [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 720.214807] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 720.214807] env[63531]: DEBUG oslo_concurrency.lockutils [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 720.218068] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-428d5973-8cef-4b98-a384-5d1dace051a6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.231990] env[63531]: DEBUG nova.compute.provider_tree [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 720.332411] env[63531]: DEBUG oslo_vmware.api [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Task: {'id': task-1117557, 'name': ReconfigVM_Task, 'duration_secs': 0.998245} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.332978] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] [instance: 7bc63fdd-4b50-486d-9681-ca0baa08f7b3] Reconfigured VM instance instance-00000022 to attach disk [datastore1] 7bc63fdd-4b50-486d-9681-ca0baa08f7b3/7bc63fdd-4b50-486d-9681-ca0baa08f7b3.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 720.333626] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ea54542a-e551-4b9c-a7b6-5c6b592cebc7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.338819] env[63531]: DEBUG oslo_vmware.api [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Waiting for the task: (returnval){ [ 720.338819] env[63531]: value = "task-1117558" [ 720.338819] env[63531]: _type = "Task" [ 720.338819] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.346443] env[63531]: DEBUG oslo_vmware.api [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Task: {'id': task-1117558, 'name': Rename_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.363703] env[63531]: DEBUG nova.network.neutron [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 720.371635] env[63531]: DEBUG oslo_vmware.api [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]529dba0f-b92e-f327-08ad-13b9dc7ca13a, 'name': SearchDatastore_Task, 'duration_secs': 0.01734} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.371936] env[63531]: DEBUG oslo_concurrency.lockutils [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 720.372762] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] f0276e34-1de3-482a-82c7-1439d40fd85e/f0276e34-1de3-482a-82c7-1439d40fd85e.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 720.372762] env[63531]: DEBUG oslo_concurrency.lockutils [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 720.372762] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 720.373172] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-08daabe3-fdf9-458d-84fd-788c50a633ec {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.376542] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a7002b7a-280f-489e-ab92-8139be93910c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.384534] env[63531]: DEBUG oslo_vmware.api [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Waiting for the task: (returnval){ [ 720.384534] env[63531]: value = "task-1117559" [ 720.384534] env[63531]: _type = "Task" [ 720.384534] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.385144] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 720.385324] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 720.389224] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b4c88cf3-52f3-407f-80ef-3858c9f338fe {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.395789] env[63531]: DEBUG oslo_vmware.api [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Task: {'id': task-1117559, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.396962] env[63531]: DEBUG oslo_vmware.api [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Waiting for the task: (returnval){ [ 720.396962] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5255882a-18f0-d7ae-d4c6-6dfb7eaa556c" [ 720.396962] env[63531]: _type = "Task" [ 720.396962] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.405723] env[63531]: DEBUG oslo_vmware.api [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5255882a-18f0-d7ae-d4c6-6dfb7eaa556c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.477472] env[63531]: DEBUG nova.compute.manager [req-aa5a841e-c055-4014-9d1a-a1521535d29a req-6c883406-5c36-4964-af8c-7e18baa2ccae service nova] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Received event network-vif-plugged-0c5ad75f-1b78-47fd-bd2e-331a7ae0732a {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 720.477692] env[63531]: DEBUG oslo_concurrency.lockutils [req-aa5a841e-c055-4014-9d1a-a1521535d29a req-6c883406-5c36-4964-af8c-7e18baa2ccae service nova] Acquiring lock "c3d597b4-1dc3-48d8-9bee-e73c8929181b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 720.477906] env[63531]: DEBUG oslo_concurrency.lockutils [req-aa5a841e-c055-4014-9d1a-a1521535d29a req-6c883406-5c36-4964-af8c-7e18baa2ccae service nova] Lock "c3d597b4-1dc3-48d8-9bee-e73c8929181b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 720.478080] env[63531]: DEBUG oslo_concurrency.lockutils [req-aa5a841e-c055-4014-9d1a-a1521535d29a req-6c883406-5c36-4964-af8c-7e18baa2ccae service nova] Lock "c3d597b4-1dc3-48d8-9bee-e73c8929181b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 720.478251] env[63531]: DEBUG nova.compute.manager [req-aa5a841e-c055-4014-9d1a-a1521535d29a req-6c883406-5c36-4964-af8c-7e18baa2ccae service nova] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] No waiting events found dispatching network-vif-plugged-0c5ad75f-1b78-47fd-bd2e-331a7ae0732a {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 720.478449] env[63531]: WARNING nova.compute.manager [req-aa5a841e-c055-4014-9d1a-a1521535d29a req-6c883406-5c36-4964-af8c-7e18baa2ccae service nova] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Received unexpected event network-vif-plugged-0c5ad75f-1b78-47fd-bd2e-331a7ae0732a for instance with vm_state building and task_state spawning. [ 720.478564] env[63531]: DEBUG nova.compute.manager [req-aa5a841e-c055-4014-9d1a-a1521535d29a req-6c883406-5c36-4964-af8c-7e18baa2ccae service nova] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Received event network-changed-0c5ad75f-1b78-47fd-bd2e-331a7ae0732a {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 720.478714] env[63531]: DEBUG nova.compute.manager [req-aa5a841e-c055-4014-9d1a-a1521535d29a req-6c883406-5c36-4964-af8c-7e18baa2ccae service nova] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Refreshing instance network info cache due to event network-changed-0c5ad75f-1b78-47fd-bd2e-331a7ae0732a. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 720.478914] env[63531]: DEBUG oslo_concurrency.lockutils [req-aa5a841e-c055-4014-9d1a-a1521535d29a req-6c883406-5c36-4964-af8c-7e18baa2ccae service nova] Acquiring lock "refresh_cache-c3d597b4-1dc3-48d8-9bee-e73c8929181b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 720.512554] env[63531]: DEBUG nova.network.neutron [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: 3f3113dd-863f-442d-a214-28dde377d227] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.521493] env[63531]: DEBUG nova.network.neutron [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Updating instance_info_cache with network_info: [{"id": "0c5ad75f-1b78-47fd-bd2e-331a7ae0732a", "address": "fa:16:3e:e6:7f:b5", "network": {"id": "c3cf494e-23bd-480f-85d0-918fff65c5fa", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1373813036-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e2efa9eb41854d9cbcf96ebb8edbcfc9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4db2961d-273d-4634-9d06-a94fa9d384fb", "external-id": "nsx-vlan-transportzone-572", "segmentation_id": 572, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c5ad75f-1b", "ovs_interfaceid": "0c5ad75f-1b78-47fd-bd2e-331a7ae0732a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.735654] env[63531]: DEBUG nova.scheduler.client.report [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 720.850158] env[63531]: DEBUG oslo_vmware.api [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Task: {'id': task-1117558, 'name': Rename_Task, 'duration_secs': 0.140894} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.850427] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] [instance: 7bc63fdd-4b50-486d-9681-ca0baa08f7b3] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 720.850649] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1fe3cc8b-f607-4bd7-bbf7-e6c6d2fc6536 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.856340] env[63531]: DEBUG oslo_vmware.api [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Waiting for the task: (returnval){ [ 720.856340] env[63531]: value = "task-1117560" [ 720.856340] env[63531]: _type = "Task" [ 720.856340] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.864669] env[63531]: DEBUG oslo_vmware.api [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Task: {'id': task-1117560, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.893952] env[63531]: DEBUG oslo_vmware.api [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Task: {'id': task-1117559, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.465003} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.894231] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] f0276e34-1de3-482a-82c7-1439d40fd85e/f0276e34-1de3-482a-82c7-1439d40fd85e.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 720.894447] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] [instance: f0276e34-1de3-482a-82c7-1439d40fd85e] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 720.894701] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-60ae1158-1c48-4f9e-9ef0-cae48ca90791 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.901952] env[63531]: DEBUG oslo_vmware.api [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Waiting for the task: (returnval){ [ 720.901952] env[63531]: value = "task-1117561" [ 720.901952] env[63531]: _type = "Task" [ 720.901952] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.908669] env[63531]: DEBUG oslo_vmware.api [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5255882a-18f0-d7ae-d4c6-6dfb7eaa556c, 'name': SearchDatastore_Task, 'duration_secs': 0.010955} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.909887] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-22ff3876-00f8-4be6-a53c-c5bb248c7082 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.915049] env[63531]: DEBUG oslo_vmware.api [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Task: {'id': task-1117561, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.918236] env[63531]: DEBUG oslo_vmware.api [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Waiting for the task: (returnval){ [ 720.918236] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]520c1ab4-447b-41e1-fe08-078ed4398f22" [ 720.918236] env[63531]: _type = "Task" [ 720.918236] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.925693] env[63531]: DEBUG oslo_vmware.api [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]520c1ab4-447b-41e1-fe08-078ed4398f22, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.015688] env[63531]: INFO nova.compute.manager [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: 3f3113dd-863f-442d-a214-28dde377d227] Took 1.02 seconds to deallocate network for instance. [ 721.024097] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Releasing lock "refresh_cache-c3d597b4-1dc3-48d8-9bee-e73c8929181b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 721.024411] env[63531]: DEBUG nova.compute.manager [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Instance network_info: |[{"id": "0c5ad75f-1b78-47fd-bd2e-331a7ae0732a", "address": "fa:16:3e:e6:7f:b5", "network": {"id": "c3cf494e-23bd-480f-85d0-918fff65c5fa", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1373813036-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e2efa9eb41854d9cbcf96ebb8edbcfc9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4db2961d-273d-4634-9d06-a94fa9d384fb", "external-id": "nsx-vlan-transportzone-572", "segmentation_id": 572, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c5ad75f-1b", "ovs_interfaceid": "0c5ad75f-1b78-47fd-bd2e-331a7ae0732a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 721.024710] env[63531]: DEBUG oslo_concurrency.lockutils [req-aa5a841e-c055-4014-9d1a-a1521535d29a req-6c883406-5c36-4964-af8c-7e18baa2ccae service nova] Acquired lock "refresh_cache-c3d597b4-1dc3-48d8-9bee-e73c8929181b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.024894] env[63531]: DEBUG nova.network.neutron [req-aa5a841e-c055-4014-9d1a-a1521535d29a req-6c883406-5c36-4964-af8c-7e18baa2ccae service nova] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Refreshing network info cache for port 0c5ad75f-1b78-47fd-bd2e-331a7ae0732a {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 721.025930] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e6:7f:b5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4db2961d-273d-4634-9d06-a94fa9d384fb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0c5ad75f-1b78-47fd-bd2e-331a7ae0732a', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 721.033547] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Creating folder: Project (e2efa9eb41854d9cbcf96ebb8edbcfc9). Parent ref: group-v244585. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 721.034280] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-79804ce9-c723-4d4f-bd2b-ac2ada3cc925 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.045820] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Created folder: Project (e2efa9eb41854d9cbcf96ebb8edbcfc9) in parent group-v244585. [ 721.046044] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Creating folder: Instances. Parent ref: group-v244615. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 721.046318] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b5c6d1c8-5205-4976-a8f9-eee37ea09202 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.055833] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Created folder: Instances in parent group-v244615. [ 721.055833] env[63531]: DEBUG oslo.service.loopingcall [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 721.055833] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 721.055833] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b792b165-0270-462e-bf63-60c7a8d78ff4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.075239] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 721.075239] env[63531]: value = "task-1117564" [ 721.075239] env[63531]: _type = "Task" [ 721.075239] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.082981] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117564, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.240946] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.397s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 721.241538] env[63531]: DEBUG nova.compute.manager [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: 645abb7a-cf73-4fb0-a9a0-49205f060d6f] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 721.244227] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.917s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 721.245732] env[63531]: INFO nova.compute.claims [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: f39c8807-1e15-41cf-899b-a1fbe0695d58] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 721.367341] env[63531]: DEBUG oslo_vmware.api [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Task: {'id': task-1117560, 'name': PowerOnVM_Task, 'duration_secs': 0.49437} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.367627] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] [instance: 7bc63fdd-4b50-486d-9681-ca0baa08f7b3] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 721.367865] env[63531]: INFO nova.compute.manager [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] [instance: 7bc63fdd-4b50-486d-9681-ca0baa08f7b3] Took 7.37 seconds to spawn the instance on the hypervisor. [ 721.368059] env[63531]: DEBUG nova.compute.manager [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] [instance: 7bc63fdd-4b50-486d-9681-ca0baa08f7b3] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 721.368881] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5114f2b-2b8e-4362-956a-9a34bc697562 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.412601] env[63531]: DEBUG oslo_vmware.api [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Task: {'id': task-1117561, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060156} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.412664] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] [instance: f0276e34-1de3-482a-82c7-1439d40fd85e] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 721.413451] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-470a3d62-27a3-4a0e-9967-8747079758f8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.437191] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] [instance: f0276e34-1de3-482a-82c7-1439d40fd85e] Reconfiguring VM instance instance-00000023 to attach disk [datastore1] f0276e34-1de3-482a-82c7-1439d40fd85e/f0276e34-1de3-482a-82c7-1439d40fd85e.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 721.440673] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c0f91c6b-1c9a-4bc2-9bbe-b2658c427126 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.460662] env[63531]: DEBUG oslo_vmware.api [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]520c1ab4-447b-41e1-fe08-078ed4398f22, 'name': SearchDatastore_Task, 'duration_secs': 0.008019} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.461927] env[63531]: DEBUG oslo_concurrency.lockutils [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 721.462252] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 0c948655-362d-4673-b3e6-3f28ec69ea3c/0c948655-362d-4673-b3e6-3f28ec69ea3c.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 721.462533] env[63531]: DEBUG oslo_vmware.api [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Waiting for the task: (returnval){ [ 721.462533] env[63531]: value = "task-1117565" [ 721.462533] env[63531]: _type = "Task" [ 721.462533] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.462707] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-612aafe2-f9af-49c4-af65-57be3c5f3405 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.471576] env[63531]: DEBUG oslo_vmware.api [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Waiting for the task: (returnval){ [ 721.471576] env[63531]: value = "task-1117566" [ 721.471576] env[63531]: _type = "Task" [ 721.471576] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.474376] env[63531]: DEBUG oslo_vmware.api [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Task: {'id': task-1117565, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.587025] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117564, 'name': CreateVM_Task, 'duration_secs': 0.322299} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.588348] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 721.588348] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 721.588348] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.588580] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 721.588720] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7587fe81-75e9-4aa8-a955-a587ec319334 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.593372] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Waiting for the task: (returnval){ [ 721.593372] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5262d734-1d5b-f538-14d6-d98fe9211d55" [ 721.593372] env[63531]: _type = "Task" [ 721.593372] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.601150] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5262d734-1d5b-f538-14d6-d98fe9211d55, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.738038] env[63531]: DEBUG nova.network.neutron [req-aa5a841e-c055-4014-9d1a-a1521535d29a req-6c883406-5c36-4964-af8c-7e18baa2ccae service nova] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Updated VIF entry in instance network info cache for port 0c5ad75f-1b78-47fd-bd2e-331a7ae0732a. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 721.738455] env[63531]: DEBUG nova.network.neutron [req-aa5a841e-c055-4014-9d1a-a1521535d29a req-6c883406-5c36-4964-af8c-7e18baa2ccae service nova] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Updating instance_info_cache with network_info: [{"id": "0c5ad75f-1b78-47fd-bd2e-331a7ae0732a", "address": "fa:16:3e:e6:7f:b5", "network": {"id": "c3cf494e-23bd-480f-85d0-918fff65c5fa", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1373813036-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e2efa9eb41854d9cbcf96ebb8edbcfc9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4db2961d-273d-4634-9d06-a94fa9d384fb", "external-id": "nsx-vlan-transportzone-572", "segmentation_id": 572, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c5ad75f-1b", "ovs_interfaceid": "0c5ad75f-1b78-47fd-bd2e-331a7ae0732a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.753902] env[63531]: DEBUG nova.compute.utils [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 721.757197] env[63531]: DEBUG nova.compute.manager [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: 645abb7a-cf73-4fb0-a9a0-49205f060d6f] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 721.757289] env[63531]: DEBUG nova.network.neutron [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: 645abb7a-cf73-4fb0-a9a0-49205f060d6f] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 721.799634] env[63531]: DEBUG nova.policy [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '43a22fda8420440eb0e60eab86eaed1a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e2efa9eb41854d9cbcf96ebb8edbcfc9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 721.890095] env[63531]: INFO nova.compute.manager [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] [instance: 7bc63fdd-4b50-486d-9681-ca0baa08f7b3] Took 41.77 seconds to build instance. [ 721.975227] env[63531]: DEBUG oslo_vmware.api [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Task: {'id': task-1117565, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.982845] env[63531]: DEBUG oslo_vmware.api [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Task: {'id': task-1117566, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.045978] env[63531]: DEBUG nova.network.neutron [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: 645abb7a-cf73-4fb0-a9a0-49205f060d6f] Successfully created port: 7d4594d0-2a5d-4a5c-9f4c-61bbe9731787 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 722.048810] env[63531]: INFO nova.scheduler.client.report [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Deleted allocations for instance 3f3113dd-863f-442d-a214-28dde377d227 [ 722.106255] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5262d734-1d5b-f538-14d6-d98fe9211d55, 'name': SearchDatastore_Task, 'duration_secs': 0.009726} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.106569] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 722.106833] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 722.107092] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 722.107240] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 722.107418] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 722.107767] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f4a17c24-d574-4483-a1df-f62b1c1e058e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.117302] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 722.117467] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 722.118575] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d0c162c2-20f1-4021-82a8-79bb90c83f32 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.123342] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Waiting for the task: (returnval){ [ 722.123342] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52dc4bd9-cf89-35ff-200d-8d18f76d3043" [ 722.123342] env[63531]: _type = "Task" [ 722.123342] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.131356] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52dc4bd9-cf89-35ff-200d-8d18f76d3043, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.241474] env[63531]: DEBUG oslo_concurrency.lockutils [req-aa5a841e-c055-4014-9d1a-a1521535d29a req-6c883406-5c36-4964-af8c-7e18baa2ccae service nova] Releasing lock "refresh_cache-c3d597b4-1dc3-48d8-9bee-e73c8929181b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 722.259245] env[63531]: DEBUG nova.compute.manager [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: 645abb7a-cf73-4fb0-a9a0-49205f060d6f] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 722.390792] env[63531]: DEBUG oslo_concurrency.lockutils [None req-592a8cee-9901-45ca-b5af-d58d98efb57f tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Lock "7bc63fdd-4b50-486d-9681-ca0baa08f7b3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 118.799s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 722.474771] env[63531]: DEBUG oslo_vmware.api [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Task: {'id': task-1117565, 'name': ReconfigVM_Task, 'duration_secs': 0.840923} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.477300] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] [instance: f0276e34-1de3-482a-82c7-1439d40fd85e] Reconfigured VM instance instance-00000023 to attach disk [datastore1] f0276e34-1de3-482a-82c7-1439d40fd85e/f0276e34-1de3-482a-82c7-1439d40fd85e.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 722.478528] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c465d07c-46b9-4465-a356-2dbc5d1c687d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.489029] env[63531]: DEBUG oslo_vmware.api [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Task: {'id': task-1117566, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.920564} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.492472] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 0c948655-362d-4673-b3e6-3f28ec69ea3c/0c948655-362d-4673-b3e6-3f28ec69ea3c.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 722.492797] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 722.493207] env[63531]: DEBUG oslo_vmware.api [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Waiting for the task: (returnval){ [ 722.493207] env[63531]: value = "task-1117567" [ 722.493207] env[63531]: _type = "Task" [ 722.493207] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.494070] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-06d33905-d92a-4adc-b748-ccb2af31e634 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.505508] env[63531]: DEBUG oslo_vmware.api [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Task: {'id': task-1117567, 'name': Rename_Task} progress is 10%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.506369] env[63531]: DEBUG oslo_vmware.api [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Waiting for the task: (returnval){ [ 722.506369] env[63531]: value = "task-1117568" [ 722.506369] env[63531]: _type = "Task" [ 722.506369] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.516155] env[63531]: DEBUG oslo_vmware.api [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Task: {'id': task-1117568, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.559685] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b496c4d8-5312-4a24-89e6-7bc5ec8e56b3 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Lock "3f3113dd-863f-442d-a214-28dde377d227" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 126.011s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 722.623969] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abe5a83a-8ab5-4661-8a12-0624a7d0e43a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.635678] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52dc4bd9-cf89-35ff-200d-8d18f76d3043, 'name': SearchDatastore_Task, 'duration_secs': 0.018011} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.638210] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b4df3b5d-3ecf-4844-8201-aa468649a8e9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.641181] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74445367-71c9-42d5-bb57-b5e967cd489c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.647421] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Waiting for the task: (returnval){ [ 722.647421] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52bf1790-dc00-eb49-bb0a-691f5e54b14f" [ 722.647421] env[63531]: _type = "Task" [ 722.647421] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.675418] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff372e24-1ae8-4911-b0df-e2c513c0cb73 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.689560] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24b41266-2fbd-4ce5-ab85-715a867709c4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.695548] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52bf1790-dc00-eb49-bb0a-691f5e54b14f, 'name': SearchDatastore_Task, 'duration_secs': 0.010573} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.696392] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 722.696727] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] c3d597b4-1dc3-48d8-9bee-e73c8929181b/c3d597b4-1dc3-48d8-9bee-e73c8929181b.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 722.697348] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9375040e-8c93-41f2-bd55-b714272030d8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.708017] env[63531]: DEBUG nova.compute.provider_tree [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 722.715111] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Waiting for the task: (returnval){ [ 722.715111] env[63531]: value = "task-1117569" [ 722.715111] env[63531]: _type = "Task" [ 722.715111] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.722357] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': task-1117569, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.897210] env[63531]: DEBUG nova.compute.manager [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 723.007191] env[63531]: DEBUG oslo_vmware.api [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Task: {'id': task-1117567, 'name': Rename_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.015869] env[63531]: DEBUG oslo_vmware.api [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Task: {'id': task-1117568, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.151707} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.016355] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 723.017738] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e960179a-0ac3-4ec4-be8b-99367ec9cbd5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.040763] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Reconfiguring VM instance instance-00000021 to attach disk [datastore1] 0c948655-362d-4673-b3e6-3f28ec69ea3c/0c948655-362d-4673-b3e6-3f28ec69ea3c.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 723.041140] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d349e651-4fb0-42b5-9129-da8ccc83af35 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.066733] env[63531]: DEBUG nova.compute.manager [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] [instance: 0af99240-fa7a-4eed-a729-a5ae98c41cf9] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 723.070302] env[63531]: DEBUG oslo_vmware.api [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Waiting for the task: (returnval){ [ 723.070302] env[63531]: value = "task-1117570" [ 723.070302] env[63531]: _type = "Task" [ 723.070302] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.086961] env[63531]: DEBUG oslo_vmware.api [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Task: {'id': task-1117570, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.183528] env[63531]: DEBUG oslo_concurrency.lockutils [None req-835ac754-8e34-4ef9-903d-b2d75006b444 tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Acquiring lock "7bc63fdd-4b50-486d-9681-ca0baa08f7b3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 723.183528] env[63531]: DEBUG oslo_concurrency.lockutils [None req-835ac754-8e34-4ef9-903d-b2d75006b444 tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Lock "7bc63fdd-4b50-486d-9681-ca0baa08f7b3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 723.183705] env[63531]: DEBUG oslo_concurrency.lockutils [None req-835ac754-8e34-4ef9-903d-b2d75006b444 tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Acquiring lock "7bc63fdd-4b50-486d-9681-ca0baa08f7b3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 723.184620] env[63531]: DEBUG oslo_concurrency.lockutils [None req-835ac754-8e34-4ef9-903d-b2d75006b444 tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Lock "7bc63fdd-4b50-486d-9681-ca0baa08f7b3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 723.184620] env[63531]: DEBUG oslo_concurrency.lockutils [None req-835ac754-8e34-4ef9-903d-b2d75006b444 tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Lock "7bc63fdd-4b50-486d-9681-ca0baa08f7b3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.186569] env[63531]: INFO nova.compute.manager [None req-835ac754-8e34-4ef9-903d-b2d75006b444 tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] [instance: 7bc63fdd-4b50-486d-9681-ca0baa08f7b3] Terminating instance [ 723.188520] env[63531]: DEBUG nova.compute.manager [None req-835ac754-8e34-4ef9-903d-b2d75006b444 tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] [instance: 7bc63fdd-4b50-486d-9681-ca0baa08f7b3] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 723.188520] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-835ac754-8e34-4ef9-903d-b2d75006b444 tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] [instance: 7bc63fdd-4b50-486d-9681-ca0baa08f7b3] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 723.189292] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-095130fe-7b72-42ca-80fa-b505571aced8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.198466] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-835ac754-8e34-4ef9-903d-b2d75006b444 tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] [instance: 7bc63fdd-4b50-486d-9681-ca0baa08f7b3] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 723.198679] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-49034822-967a-453e-9a9b-3a9aece9c5df {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.204883] env[63531]: DEBUG oslo_vmware.api [None req-835ac754-8e34-4ef9-903d-b2d75006b444 tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Waiting for the task: (returnval){ [ 723.204883] env[63531]: value = "task-1117571" [ 723.204883] env[63531]: _type = "Task" [ 723.204883] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.212847] env[63531]: DEBUG nova.scheduler.client.report [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 723.215698] env[63531]: DEBUG oslo_vmware.api [None req-835ac754-8e34-4ef9-903d-b2d75006b444 tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Task: {'id': task-1117571, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.223522] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': task-1117569, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.472028} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.223745] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] c3d597b4-1dc3-48d8-9bee-e73c8929181b/c3d597b4-1dc3-48d8-9bee-e73c8929181b.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 723.223946] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 723.224181] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9dff2377-4ee8-46a7-bf7a-6a8ea1b4b04c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.230040] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Waiting for the task: (returnval){ [ 723.230040] env[63531]: value = "task-1117572" [ 723.230040] env[63531]: _type = "Task" [ 723.230040] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.238382] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': task-1117572, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.271665] env[63531]: DEBUG nova.compute.manager [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: 645abb7a-cf73-4fb0-a9a0-49205f060d6f] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 723.298272] env[63531]: DEBUG nova.virt.hardware [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 723.298520] env[63531]: DEBUG nova.virt.hardware [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 723.298736] env[63531]: DEBUG nova.virt.hardware [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 723.298992] env[63531]: DEBUG nova.virt.hardware [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 723.299112] env[63531]: DEBUG nova.virt.hardware [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 723.299260] env[63531]: DEBUG nova.virt.hardware [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 723.299465] env[63531]: DEBUG nova.virt.hardware [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 723.299640] env[63531]: DEBUG nova.virt.hardware [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 723.299781] env[63531]: DEBUG nova.virt.hardware [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 723.299941] env[63531]: DEBUG nova.virt.hardware [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 723.300135] env[63531]: DEBUG nova.virt.hardware [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 723.301286] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-429a15f2-d3e4-4762-8956-d29cb01af53e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.309374] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c5f04aa-dfff-4c78-ae2f-ae0f32a46963 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.421352] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 723.506507] env[63531]: DEBUG oslo_vmware.api [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Task: {'id': task-1117567, 'name': Rename_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.549859] env[63531]: DEBUG nova.compute.manager [req-ef047ef1-d09c-41eb-9a02-1b65d9329bf5 req-49fcb014-3fed-445c-adf1-6c629879147f service nova] [instance: 645abb7a-cf73-4fb0-a9a0-49205f060d6f] Received event network-vif-plugged-7d4594d0-2a5d-4a5c-9f4c-61bbe9731787 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 723.550163] env[63531]: DEBUG oslo_concurrency.lockutils [req-ef047ef1-d09c-41eb-9a02-1b65d9329bf5 req-49fcb014-3fed-445c-adf1-6c629879147f service nova] Acquiring lock "645abb7a-cf73-4fb0-a9a0-49205f060d6f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 723.550317] env[63531]: DEBUG oslo_concurrency.lockutils [req-ef047ef1-d09c-41eb-9a02-1b65d9329bf5 req-49fcb014-3fed-445c-adf1-6c629879147f service nova] Lock "645abb7a-cf73-4fb0-a9a0-49205f060d6f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 723.550452] env[63531]: DEBUG oslo_concurrency.lockutils [req-ef047ef1-d09c-41eb-9a02-1b65d9329bf5 req-49fcb014-3fed-445c-adf1-6c629879147f service nova] Lock "645abb7a-cf73-4fb0-a9a0-49205f060d6f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.550618] env[63531]: DEBUG nova.compute.manager [req-ef047ef1-d09c-41eb-9a02-1b65d9329bf5 req-49fcb014-3fed-445c-adf1-6c629879147f service nova] [instance: 645abb7a-cf73-4fb0-a9a0-49205f060d6f] No waiting events found dispatching network-vif-plugged-7d4594d0-2a5d-4a5c-9f4c-61bbe9731787 {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 723.550780] env[63531]: WARNING nova.compute.manager [req-ef047ef1-d09c-41eb-9a02-1b65d9329bf5 req-49fcb014-3fed-445c-adf1-6c629879147f service nova] [instance: 645abb7a-cf73-4fb0-a9a0-49205f060d6f] Received unexpected event network-vif-plugged-7d4594d0-2a5d-4a5c-9f4c-61bbe9731787 for instance with vm_state building and task_state spawning. [ 723.562332] env[63531]: DEBUG oslo_concurrency.lockutils [None req-396bcb9b-cfea-4ccb-9d40-6e75c674fc51 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Acquiring lock "ffb34f90-d22f-440b-ba07-75d474c3c300" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 723.562553] env[63531]: DEBUG oslo_concurrency.lockutils [None req-396bcb9b-cfea-4ccb-9d40-6e75c674fc51 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Lock "ffb34f90-d22f-440b-ba07-75d474c3c300" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 723.562747] env[63531]: DEBUG oslo_concurrency.lockutils [None req-396bcb9b-cfea-4ccb-9d40-6e75c674fc51 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Acquiring lock "ffb34f90-d22f-440b-ba07-75d474c3c300-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 723.562925] env[63531]: DEBUG oslo_concurrency.lockutils [None req-396bcb9b-cfea-4ccb-9d40-6e75c674fc51 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Lock "ffb34f90-d22f-440b-ba07-75d474c3c300-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 723.563099] env[63531]: DEBUG oslo_concurrency.lockutils [None req-396bcb9b-cfea-4ccb-9d40-6e75c674fc51 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Lock "ffb34f90-d22f-440b-ba07-75d474c3c300-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.565211] env[63531]: INFO nova.compute.manager [None req-396bcb9b-cfea-4ccb-9d40-6e75c674fc51 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: ffb34f90-d22f-440b-ba07-75d474c3c300] Terminating instance [ 723.568265] env[63531]: DEBUG nova.compute.manager [None req-396bcb9b-cfea-4ccb-9d40-6e75c674fc51 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: ffb34f90-d22f-440b-ba07-75d474c3c300] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 723.568265] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-396bcb9b-cfea-4ccb-9d40-6e75c674fc51 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: ffb34f90-d22f-440b-ba07-75d474c3c300] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 723.569149] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f6337fc-84a2-4c2a-b2c2-49209918bde9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.581232] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-396bcb9b-cfea-4ccb-9d40-6e75c674fc51 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: ffb34f90-d22f-440b-ba07-75d474c3c300] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 723.584266] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dc437243-bf4a-46f3-9fa9-ebeff05dee78 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.586023] env[63531]: DEBUG oslo_vmware.api [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Task: {'id': task-1117570, 'name': ReconfigVM_Task, 'duration_secs': 0.36685} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.586277] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Reconfigured VM instance instance-00000021 to attach disk [datastore1] 0c948655-362d-4673-b3e6-3f28ec69ea3c/0c948655-362d-4673-b3e6-3f28ec69ea3c.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 723.587151] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b5446514-4a70-445d-932e-ad4c701edd0c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.589172] env[63531]: DEBUG nova.network.neutron [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: 645abb7a-cf73-4fb0-a9a0-49205f060d6f] Successfully updated port: 7d4594d0-2a5d-4a5c-9f4c-61bbe9731787 {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 723.591289] env[63531]: DEBUG oslo_vmware.api [None req-396bcb9b-cfea-4ccb-9d40-6e75c674fc51 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Waiting for the task: (returnval){ [ 723.591289] env[63531]: value = "task-1117573" [ 723.591289] env[63531]: _type = "Task" [ 723.591289] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.592088] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 723.596338] env[63531]: DEBUG oslo_vmware.api [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Waiting for the task: (returnval){ [ 723.596338] env[63531]: value = "task-1117574" [ 723.596338] env[63531]: _type = "Task" [ 723.596338] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.602880] env[63531]: DEBUG oslo_vmware.api [None req-396bcb9b-cfea-4ccb-9d40-6e75c674fc51 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Task: {'id': task-1117573, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.608228] env[63531]: DEBUG oslo_vmware.api [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Task: {'id': task-1117574, 'name': Rename_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.715564] env[63531]: DEBUG oslo_vmware.api [None req-835ac754-8e34-4ef9-903d-b2d75006b444 tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Task: {'id': task-1117571, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.717620] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.473s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.718743] env[63531]: DEBUG nova.compute.manager [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: f39c8807-1e15-41cf-899b-a1fbe0695d58] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 723.721625] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.099s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 723.722841] env[63531]: INFO nova.compute.claims [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 723.740403] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': task-1117572, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066362} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.740712] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 723.741636] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91ed0b3c-7301-4048-a4d5-feff63e510d0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.765061] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Reconfiguring VM instance instance-00000024 to attach disk [datastore2] c3d597b4-1dc3-48d8-9bee-e73c8929181b/c3d597b4-1dc3-48d8-9bee-e73c8929181b.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 723.766875] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3577e272-9022-468b-ac06-b290d4659a05 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.788345] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Waiting for the task: (returnval){ [ 723.788345] env[63531]: value = "task-1117575" [ 723.788345] env[63531]: _type = "Task" [ 723.788345] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.796877] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': task-1117575, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.008186] env[63531]: DEBUG oslo_vmware.api [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Task: {'id': task-1117567, 'name': Rename_Task, 'duration_secs': 1.32321} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.008186] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] [instance: f0276e34-1de3-482a-82c7-1439d40fd85e] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 724.008487] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9accdf41-6223-4d8f-9785-ee3e3fc93804 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.015183] env[63531]: DEBUG oslo_vmware.api [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Waiting for the task: (returnval){ [ 724.015183] env[63531]: value = "task-1117576" [ 724.015183] env[63531]: _type = "Task" [ 724.015183] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.023215] env[63531]: DEBUG oslo_vmware.api [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Task: {'id': task-1117576, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.097669] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Acquiring lock "refresh_cache-645abb7a-cf73-4fb0-a9a0-49205f060d6f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 724.097985] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Acquired lock "refresh_cache-645abb7a-cf73-4fb0-a9a0-49205f060d6f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 724.098253] env[63531]: DEBUG nova.network.neutron [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: 645abb7a-cf73-4fb0-a9a0-49205f060d6f] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 724.108355] env[63531]: DEBUG oslo_vmware.api [None req-396bcb9b-cfea-4ccb-9d40-6e75c674fc51 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Task: {'id': task-1117573, 'name': PowerOffVM_Task, 'duration_secs': 0.382401} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.112193] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-396bcb9b-cfea-4ccb-9d40-6e75c674fc51 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: ffb34f90-d22f-440b-ba07-75d474c3c300] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 724.112193] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-396bcb9b-cfea-4ccb-9d40-6e75c674fc51 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: ffb34f90-d22f-440b-ba07-75d474c3c300] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 724.112632] env[63531]: DEBUG oslo_vmware.api [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Task: {'id': task-1117574, 'name': Rename_Task, 'duration_secs': 0.134475} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.112838] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d24a79e4-57d2-4dfb-a277-9b871f1cb0a0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.114361] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 724.114823] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-276b6d2e-3f2d-4268-881e-921c9bccd230 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.121305] env[63531]: DEBUG oslo_vmware.api [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Waiting for the task: (returnval){ [ 724.121305] env[63531]: value = "task-1117578" [ 724.121305] env[63531]: _type = "Task" [ 724.121305] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.128992] env[63531]: DEBUG oslo_vmware.api [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Task: {'id': task-1117578, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.187329] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-396bcb9b-cfea-4ccb-9d40-6e75c674fc51 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: ffb34f90-d22f-440b-ba07-75d474c3c300] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 724.187568] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-396bcb9b-cfea-4ccb-9d40-6e75c674fc51 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: ffb34f90-d22f-440b-ba07-75d474c3c300] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 724.188026] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-396bcb9b-cfea-4ccb-9d40-6e75c674fc51 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Deleting the datastore file [datastore1] ffb34f90-d22f-440b-ba07-75d474c3c300 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 724.188329] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bdd1beba-ad38-4794-bb99-e9dc475603e2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.195814] env[63531]: DEBUG oslo_vmware.api [None req-396bcb9b-cfea-4ccb-9d40-6e75c674fc51 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Waiting for the task: (returnval){ [ 724.195814] env[63531]: value = "task-1117579" [ 724.195814] env[63531]: _type = "Task" [ 724.195814] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.203327] env[63531]: DEBUG oslo_vmware.api [None req-396bcb9b-cfea-4ccb-9d40-6e75c674fc51 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Task: {'id': task-1117579, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.213834] env[63531]: DEBUG oslo_vmware.api [None req-835ac754-8e34-4ef9-903d-b2d75006b444 tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Task: {'id': task-1117571, 'name': PowerOffVM_Task, 'duration_secs': 0.768798} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.214093] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-835ac754-8e34-4ef9-903d-b2d75006b444 tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] [instance: 7bc63fdd-4b50-486d-9681-ca0baa08f7b3] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 724.214263] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-835ac754-8e34-4ef9-903d-b2d75006b444 tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] [instance: 7bc63fdd-4b50-486d-9681-ca0baa08f7b3] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 724.214503] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4b9cce00-12d9-40a7-8c58-b861653f9f0f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.227594] env[63531]: DEBUG nova.compute.utils [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 724.232543] env[63531]: DEBUG nova.compute.manager [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: f39c8807-1e15-41cf-899b-a1fbe0695d58] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 724.232731] env[63531]: DEBUG nova.network.neutron [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: f39c8807-1e15-41cf-899b-a1fbe0695d58] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 724.279133] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-835ac754-8e34-4ef9-903d-b2d75006b444 tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] [instance: 7bc63fdd-4b50-486d-9681-ca0baa08f7b3] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 724.279350] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-835ac754-8e34-4ef9-903d-b2d75006b444 tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] [instance: 7bc63fdd-4b50-486d-9681-ca0baa08f7b3] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 724.279529] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-835ac754-8e34-4ef9-903d-b2d75006b444 tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Deleting the datastore file [datastore1] 7bc63fdd-4b50-486d-9681-ca0baa08f7b3 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 724.279781] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-95bd8bdb-5030-4864-9eb6-12e53b7720ca {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.286535] env[63531]: DEBUG oslo_vmware.api [None req-835ac754-8e34-4ef9-903d-b2d75006b444 tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Waiting for the task: (returnval){ [ 724.286535] env[63531]: value = "task-1117581" [ 724.286535] env[63531]: _type = "Task" [ 724.286535] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.288269] env[63531]: DEBUG nova.policy [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '43a22fda8420440eb0e60eab86eaed1a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e2efa9eb41854d9cbcf96ebb8edbcfc9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 724.304054] env[63531]: DEBUG oslo_vmware.api [None req-835ac754-8e34-4ef9-903d-b2d75006b444 tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Task: {'id': task-1117581, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.304305] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': task-1117575, 'name': ReconfigVM_Task, 'duration_secs': 0.30415} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.304537] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Reconfigured VM instance instance-00000024 to attach disk [datastore2] c3d597b4-1dc3-48d8-9bee-e73c8929181b/c3d597b4-1dc3-48d8-9bee-e73c8929181b.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 724.305168] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fd232396-a658-410a-a2fb-82671b9bcb77 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.311101] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Waiting for the task: (returnval){ [ 724.311101] env[63531]: value = "task-1117582" [ 724.311101] env[63531]: _type = "Task" [ 724.311101] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.320770] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': task-1117582, 'name': Rename_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.525245] env[63531]: DEBUG oslo_vmware.api [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Task: {'id': task-1117576, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.568044] env[63531]: DEBUG nova.network.neutron [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: f39c8807-1e15-41cf-899b-a1fbe0695d58] Successfully created port: 61792095-541d-4d6e-b33c-c35d459fe790 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 724.631215] env[63531]: DEBUG oslo_vmware.api [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Task: {'id': task-1117578, 'name': PowerOnVM_Task} progress is 90%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.640860] env[63531]: DEBUG nova.network.neutron [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: 645abb7a-cf73-4fb0-a9a0-49205f060d6f] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 724.705316] env[63531]: DEBUG oslo_vmware.api [None req-396bcb9b-cfea-4ccb-9d40-6e75c674fc51 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Task: {'id': task-1117579, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.196364} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.705581] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-396bcb9b-cfea-4ccb-9d40-6e75c674fc51 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 724.705797] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-396bcb9b-cfea-4ccb-9d40-6e75c674fc51 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: ffb34f90-d22f-440b-ba07-75d474c3c300] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 724.706083] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-396bcb9b-cfea-4ccb-9d40-6e75c674fc51 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: ffb34f90-d22f-440b-ba07-75d474c3c300] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 724.706301] env[63531]: INFO nova.compute.manager [None req-396bcb9b-cfea-4ccb-9d40-6e75c674fc51 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] [instance: ffb34f90-d22f-440b-ba07-75d474c3c300] Took 1.14 seconds to destroy the instance on the hypervisor. [ 724.706564] env[63531]: DEBUG oslo.service.loopingcall [None req-396bcb9b-cfea-4ccb-9d40-6e75c674fc51 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 724.709127] env[63531]: DEBUG nova.compute.manager [-] [instance: ffb34f90-d22f-440b-ba07-75d474c3c300] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 724.709316] env[63531]: DEBUG nova.network.neutron [-] [instance: ffb34f90-d22f-440b-ba07-75d474c3c300] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 724.737022] env[63531]: DEBUG nova.compute.manager [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: f39c8807-1e15-41cf-899b-a1fbe0695d58] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 724.804212] env[63531]: DEBUG oslo_vmware.api [None req-835ac754-8e34-4ef9-903d-b2d75006b444 tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Task: {'id': task-1117581, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.164309} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.804919] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-835ac754-8e34-4ef9-903d-b2d75006b444 tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 724.805347] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-835ac754-8e34-4ef9-903d-b2d75006b444 tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] [instance: 7bc63fdd-4b50-486d-9681-ca0baa08f7b3] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 724.805723] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-835ac754-8e34-4ef9-903d-b2d75006b444 tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] [instance: 7bc63fdd-4b50-486d-9681-ca0baa08f7b3] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 724.806543] env[63531]: INFO nova.compute.manager [None req-835ac754-8e34-4ef9-903d-b2d75006b444 tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] [instance: 7bc63fdd-4b50-486d-9681-ca0baa08f7b3] Took 1.62 seconds to destroy the instance on the hypervisor. [ 724.806543] env[63531]: DEBUG oslo.service.loopingcall [None req-835ac754-8e34-4ef9-903d-b2d75006b444 tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 724.806769] env[63531]: DEBUG nova.compute.manager [-] [instance: 7bc63fdd-4b50-486d-9681-ca0baa08f7b3] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 724.806968] env[63531]: DEBUG nova.network.neutron [-] [instance: 7bc63fdd-4b50-486d-9681-ca0baa08f7b3] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 724.821671] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': task-1117582, 'name': Rename_Task, 'duration_secs': 0.140018} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.824339] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 724.824770] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-54a61fa6-01f7-432d-a158-93dae38afdd0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.836138] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Waiting for the task: (returnval){ [ 724.836138] env[63531]: value = "task-1117583" [ 724.836138] env[63531]: _type = "Task" [ 724.836138] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.844811] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': task-1117583, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.892986] env[63531]: DEBUG nova.network.neutron [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: 645abb7a-cf73-4fb0-a9a0-49205f060d6f] Updating instance_info_cache with network_info: [{"id": "7d4594d0-2a5d-4a5c-9f4c-61bbe9731787", "address": "fa:16:3e:2f:1c:f5", "network": {"id": "c3cf494e-23bd-480f-85d0-918fff65c5fa", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1373813036-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e2efa9eb41854d9cbcf96ebb8edbcfc9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4db2961d-273d-4634-9d06-a94fa9d384fb", "external-id": "nsx-vlan-transportzone-572", "segmentation_id": 572, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d4594d0-2a", "ovs_interfaceid": "7d4594d0-2a5d-4a5c-9f4c-61bbe9731787", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.028933] env[63531]: DEBUG oslo_vmware.api [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Task: {'id': task-1117576, 'name': PowerOnVM_Task, 'duration_secs': 0.572306} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.032082] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] [instance: f0276e34-1de3-482a-82c7-1439d40fd85e] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 725.032302] env[63531]: INFO nova.compute.manager [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] [instance: f0276e34-1de3-482a-82c7-1439d40fd85e] Took 8.55 seconds to spawn the instance on the hypervisor. [ 725.032977] env[63531]: DEBUG nova.compute.manager [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] [instance: f0276e34-1de3-482a-82c7-1439d40fd85e] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 725.034117] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b90052fc-88ef-44c8-97b9-37772286b1b9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.133476] env[63531]: DEBUG oslo_vmware.api [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Task: {'id': task-1117578, 'name': PowerOnVM_Task, 'duration_secs': 0.559181} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.136537] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 725.136778] env[63531]: DEBUG nova.compute.manager [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 725.137773] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-969deb4d-691e-458c-9027-ae547033d844 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.185570] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d37cac43-2fd4-4fb2-b147-579b5c088c06 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.194262] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9325f73d-9cc2-4e3d-95cc-e93c6b04a429 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.232019] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55842ccd-1181-4fe2-bd9c-cef2157a2391 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.240476] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61e7316e-a588-4b42-af81-ef9f3cd70f1b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.261744] env[63531]: DEBUG nova.compute.provider_tree [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 725.344585] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': task-1117583, 'name': PowerOnVM_Task, 'duration_secs': 0.455899} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.344956] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 725.345226] env[63531]: INFO nova.compute.manager [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Took 6.41 seconds to spawn the instance on the hypervisor. [ 725.345438] env[63531]: DEBUG nova.compute.manager [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 725.346418] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-001ad042-2758-4ea2-adc2-5c1b06c5439a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.398740] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Releasing lock "refresh_cache-645abb7a-cf73-4fb0-a9a0-49205f060d6f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 725.399057] env[63531]: DEBUG nova.compute.manager [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: 645abb7a-cf73-4fb0-a9a0-49205f060d6f] Instance network_info: |[{"id": "7d4594d0-2a5d-4a5c-9f4c-61bbe9731787", "address": "fa:16:3e:2f:1c:f5", "network": {"id": "c3cf494e-23bd-480f-85d0-918fff65c5fa", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1373813036-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e2efa9eb41854d9cbcf96ebb8edbcfc9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4db2961d-273d-4634-9d06-a94fa9d384fb", "external-id": "nsx-vlan-transportzone-572", "segmentation_id": 572, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d4594d0-2a", "ovs_interfaceid": "7d4594d0-2a5d-4a5c-9f4c-61bbe9731787", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 725.399815] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: 645abb7a-cf73-4fb0-a9a0-49205f060d6f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2f:1c:f5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4db2961d-273d-4634-9d06-a94fa9d384fb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7d4594d0-2a5d-4a5c-9f4c-61bbe9731787', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 725.407295] env[63531]: DEBUG oslo.service.loopingcall [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 725.408021] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 645abb7a-cf73-4fb0-a9a0-49205f060d6f] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 725.408258] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c9a895f5-5f71-49d2-a3aa-5d7983ccda81 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.429552] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 725.429552] env[63531]: value = "task-1117584" [ 725.429552] env[63531]: _type = "Task" [ 725.429552] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.432075] env[63531]: DEBUG nova.compute.manager [req-d2482731-6596-474e-a15a-462b74ef1b4f req-4a5f5b93-c282-4acd-8be6-b98e1eabe332 service nova] [instance: ffb34f90-d22f-440b-ba07-75d474c3c300] Received event network-vif-deleted-17d053f9-c98c-4b3d-8a11-1c514de74d81 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 725.432271] env[63531]: INFO nova.compute.manager [req-d2482731-6596-474e-a15a-462b74ef1b4f req-4a5f5b93-c282-4acd-8be6-b98e1eabe332 service nova] [instance: ffb34f90-d22f-440b-ba07-75d474c3c300] Neutron deleted interface 17d053f9-c98c-4b3d-8a11-1c514de74d81; detaching it from the instance and deleting it from the info cache [ 725.432436] env[63531]: DEBUG nova.network.neutron [req-d2482731-6596-474e-a15a-462b74ef1b4f req-4a5f5b93-c282-4acd-8be6-b98e1eabe332 service nova] [instance: ffb34f90-d22f-440b-ba07-75d474c3c300] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.443585] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117584, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.556014] env[63531]: INFO nova.compute.manager [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] [instance: f0276e34-1de3-482a-82c7-1439d40fd85e] Took 43.32 seconds to build instance. [ 725.653812] env[63531]: DEBUG oslo_concurrency.lockutils [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 725.690632] env[63531]: DEBUG nova.compute.manager [req-5dfab5d7-4764-4f0d-bda1-224d1de78753 req-c66cb1d0-e79f-4479-aa53-29b955fdebc2 service nova] [instance: 645abb7a-cf73-4fb0-a9a0-49205f060d6f] Received event network-changed-7d4594d0-2a5d-4a5c-9f4c-61bbe9731787 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 725.690783] env[63531]: DEBUG nova.compute.manager [req-5dfab5d7-4764-4f0d-bda1-224d1de78753 req-c66cb1d0-e79f-4479-aa53-29b955fdebc2 service nova] [instance: 645abb7a-cf73-4fb0-a9a0-49205f060d6f] Refreshing instance network info cache due to event network-changed-7d4594d0-2a5d-4a5c-9f4c-61bbe9731787. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 725.690999] env[63531]: DEBUG oslo_concurrency.lockutils [req-5dfab5d7-4764-4f0d-bda1-224d1de78753 req-c66cb1d0-e79f-4479-aa53-29b955fdebc2 service nova] Acquiring lock "refresh_cache-645abb7a-cf73-4fb0-a9a0-49205f060d6f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 725.691165] env[63531]: DEBUG oslo_concurrency.lockutils [req-5dfab5d7-4764-4f0d-bda1-224d1de78753 req-c66cb1d0-e79f-4479-aa53-29b955fdebc2 service nova] Acquired lock "refresh_cache-645abb7a-cf73-4fb0-a9a0-49205f060d6f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.691326] env[63531]: DEBUG nova.network.neutron [req-5dfab5d7-4764-4f0d-bda1-224d1de78753 req-c66cb1d0-e79f-4479-aa53-29b955fdebc2 service nova] [instance: 645abb7a-cf73-4fb0-a9a0-49205f060d6f] Refreshing network info cache for port 7d4594d0-2a5d-4a5c-9f4c-61bbe9731787 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 725.752503] env[63531]: DEBUG nova.compute.manager [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: f39c8807-1e15-41cf-899b-a1fbe0695d58] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 725.765787] env[63531]: DEBUG nova.scheduler.client.report [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 725.777328] env[63531]: DEBUG nova.network.neutron [-] [instance: ffb34f90-d22f-440b-ba07-75d474c3c300] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.780813] env[63531]: DEBUG nova.virt.hardware [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 725.781414] env[63531]: DEBUG nova.virt.hardware [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 725.781414] env[63531]: DEBUG nova.virt.hardware [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 725.782287] env[63531]: DEBUG nova.virt.hardware [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 725.782287] env[63531]: DEBUG nova.virt.hardware [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 725.782287] env[63531]: DEBUG nova.virt.hardware [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 725.782287] env[63531]: DEBUG nova.virt.hardware [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 725.782287] env[63531]: DEBUG nova.virt.hardware [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 725.782571] env[63531]: DEBUG nova.virt.hardware [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 725.782673] env[63531]: DEBUG nova.virt.hardware [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 725.782866] env[63531]: DEBUG nova.virt.hardware [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 725.784195] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bde0d4a-dde3-4c14-9a7d-cb12e56248d9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.787892] env[63531]: INFO nova.compute.manager [-] [instance: ffb34f90-d22f-440b-ba07-75d474c3c300] Took 1.08 seconds to deallocate network for instance. [ 725.796636] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c41d8f3f-25bc-4811-bf18-5519963afc36 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.864191] env[63531]: INFO nova.compute.manager [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Took 41.63 seconds to build instance. [ 725.900169] env[63531]: DEBUG nova.network.neutron [-] [instance: 7bc63fdd-4b50-486d-9681-ca0baa08f7b3] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.936066] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e57508ed-6c52-4810-b877-deaeccb350e2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.944471] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117584, 'name': CreateVM_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.948408] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7f5111c-f132-4e74-a6c3-1ed24af05539 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.966081] env[63531]: DEBUG oslo_concurrency.lockutils [None req-2f7bff43-205d-4df7-a4ce-48fd13394e84 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Acquiring lock "interface-f0276e34-1de3-482a-82c7-1439d40fd85e-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 725.966081] env[63531]: DEBUG oslo_concurrency.lockutils [None req-2f7bff43-205d-4df7-a4ce-48fd13394e84 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Lock "interface-f0276e34-1de3-482a-82c7-1439d40fd85e-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 725.966267] env[63531]: DEBUG nova.objects.instance [None req-2f7bff43-205d-4df7-a4ce-48fd13394e84 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Lazy-loading 'flavor' on Instance uuid f0276e34-1de3-482a-82c7-1439d40fd85e {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 725.975407] env[63531]: DEBUG nova.compute.manager [req-d2482731-6596-474e-a15a-462b74ef1b4f req-4a5f5b93-c282-4acd-8be6-b98e1eabe332 service nova] [instance: ffb34f90-d22f-440b-ba07-75d474c3c300] Detach interface failed, port_id=17d053f9-c98c-4b3d-8a11-1c514de74d81, reason: Instance ffb34f90-d22f-440b-ba07-75d474c3c300 could not be found. {{(pid=63531) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 726.057821] env[63531]: DEBUG oslo_concurrency.lockutils [None req-13f2ca6a-bd72-4e2a-a86a-e79ad8784033 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Lock "f0276e34-1de3-482a-82c7-1439d40fd85e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 121.254s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 726.078249] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1a515fd1-1aca-4043-ae4d-827540148b79 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Acquiring lock "0c948655-362d-4673-b3e6-3f28ec69ea3c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 726.078488] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1a515fd1-1aca-4043-ae4d-827540148b79 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Lock "0c948655-362d-4673-b3e6-3f28ec69ea3c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 726.078678] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1a515fd1-1aca-4043-ae4d-827540148b79 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Acquiring lock "0c948655-362d-4673-b3e6-3f28ec69ea3c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 726.079069] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1a515fd1-1aca-4043-ae4d-827540148b79 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Lock "0c948655-362d-4673-b3e6-3f28ec69ea3c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 726.079069] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1a515fd1-1aca-4043-ae4d-827540148b79 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Lock "0c948655-362d-4673-b3e6-3f28ec69ea3c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 726.081456] env[63531]: INFO nova.compute.manager [None req-1a515fd1-1aca-4043-ae4d-827540148b79 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Terminating instance [ 726.082950] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1a515fd1-1aca-4043-ae4d-827540148b79 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Acquiring lock "refresh_cache-0c948655-362d-4673-b3e6-3f28ec69ea3c" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.083121] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1a515fd1-1aca-4043-ae4d-827540148b79 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Acquired lock "refresh_cache-0c948655-362d-4673-b3e6-3f28ec69ea3c" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.083301] env[63531]: DEBUG nova.network.neutron [None req-1a515fd1-1aca-4043-ae4d-827540148b79 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 726.209579] env[63531]: DEBUG nova.network.neutron [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: f39c8807-1e15-41cf-899b-a1fbe0695d58] Successfully updated port: 61792095-541d-4d6e-b33c-c35d459fe790 {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 726.272447] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.551s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 726.273092] env[63531]: DEBUG nova.compute.manager [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 726.275790] env[63531]: DEBUG oslo_concurrency.lockutils [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.739s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 726.278756] env[63531]: INFO nova.compute.claims [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 726.294606] env[63531]: DEBUG oslo_concurrency.lockutils [None req-396bcb9b-cfea-4ccb-9d40-6e75c674fc51 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 726.366458] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Lock "c3d597b4-1dc3-48d8-9bee-e73c8929181b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 120.254s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 726.402161] env[63531]: INFO nova.compute.manager [-] [instance: 7bc63fdd-4b50-486d-9681-ca0baa08f7b3] Took 1.60 seconds to deallocate network for instance. [ 726.450824] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117584, 'name': CreateVM_Task, 'duration_secs': 0.582995} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.450824] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 645abb7a-cf73-4fb0-a9a0-49205f060d6f] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 726.450824] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.450824] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.450824] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 726.454580] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-42afd228-509c-4cd5-9336-159a73a09a7c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.454580] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Waiting for the task: (returnval){ [ 726.454580] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]529dabe6-be85-9964-4b41-d4d2bc556b8b" [ 726.454580] env[63531]: _type = "Task" [ 726.454580] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.461742] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]529dabe6-be85-9964-4b41-d4d2bc556b8b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.477596] env[63531]: DEBUG nova.objects.instance [None req-2f7bff43-205d-4df7-a4ce-48fd13394e84 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Lazy-loading 'pci_requests' on Instance uuid f0276e34-1de3-482a-82c7-1439d40fd85e {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 726.531429] env[63531]: DEBUG nova.network.neutron [req-5dfab5d7-4764-4f0d-bda1-224d1de78753 req-c66cb1d0-e79f-4479-aa53-29b955fdebc2 service nova] [instance: 645abb7a-cf73-4fb0-a9a0-49205f060d6f] Updated VIF entry in instance network info cache for port 7d4594d0-2a5d-4a5c-9f4c-61bbe9731787. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 726.531779] env[63531]: DEBUG nova.network.neutron [req-5dfab5d7-4764-4f0d-bda1-224d1de78753 req-c66cb1d0-e79f-4479-aa53-29b955fdebc2 service nova] [instance: 645abb7a-cf73-4fb0-a9a0-49205f060d6f] Updating instance_info_cache with network_info: [{"id": "7d4594d0-2a5d-4a5c-9f4c-61bbe9731787", "address": "fa:16:3e:2f:1c:f5", "network": {"id": "c3cf494e-23bd-480f-85d0-918fff65c5fa", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1373813036-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e2efa9eb41854d9cbcf96ebb8edbcfc9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4db2961d-273d-4634-9d06-a94fa9d384fb", "external-id": "nsx-vlan-transportzone-572", "segmentation_id": 572, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d4594d0-2a", "ovs_interfaceid": "7d4594d0-2a5d-4a5c-9f4c-61bbe9731787", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.560822] env[63531]: DEBUG nova.compute.manager [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 726.599948] env[63531]: DEBUG nova.network.neutron [None req-1a515fd1-1aca-4043-ae4d-827540148b79 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 726.653300] env[63531]: DEBUG nova.network.neutron [None req-1a515fd1-1aca-4043-ae4d-827540148b79 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.712373] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Acquiring lock "refresh_cache-f39c8807-1e15-41cf-899b-a1fbe0695d58" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.712373] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Acquired lock "refresh_cache-f39c8807-1e15-41cf-899b-a1fbe0695d58" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.712561] env[63531]: DEBUG nova.network.neutron [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: f39c8807-1e15-41cf-899b-a1fbe0695d58] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 726.781817] env[63531]: DEBUG nova.compute.utils [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 726.785831] env[63531]: DEBUG nova.compute.manager [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 726.785950] env[63531]: DEBUG nova.network.neutron [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 726.844554] env[63531]: DEBUG nova.policy [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2e047a775d4642dca798c5d624ccbc89', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5f0d10f679de4fa4bd9d725951e536c1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 726.869130] env[63531]: DEBUG nova.compute.manager [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 726.909564] env[63531]: DEBUG oslo_concurrency.lockutils [None req-835ac754-8e34-4ef9-903d-b2d75006b444 tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 726.967614] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]529dabe6-be85-9964-4b41-d4d2bc556b8b, 'name': SearchDatastore_Task, 'duration_secs': 0.008672} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.967972] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 726.968251] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: 645abb7a-cf73-4fb0-a9a0-49205f060d6f] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 726.968493] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.968637] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.968820] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 726.969092] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9229da11-cc3e-4951-bdc3-c0b07d8af630 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.977909] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 726.978106] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 726.978799] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5cc0f7be-4f65-4039-bfeb-c53727fc344a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.981315] env[63531]: DEBUG nova.objects.base [None req-2f7bff43-205d-4df7-a4ce-48fd13394e84 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=63531) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 726.981508] env[63531]: DEBUG nova.network.neutron [None req-2f7bff43-205d-4df7-a4ce-48fd13394e84 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] [instance: f0276e34-1de3-482a-82c7-1439d40fd85e] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 726.985851] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Waiting for the task: (returnval){ [ 726.985851] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52a4b365-0d22-556a-b6d9-283953a57539" [ 726.985851] env[63531]: _type = "Task" [ 726.985851] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.994483] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52a4b365-0d22-556a-b6d9-283953a57539, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.034325] env[63531]: DEBUG oslo_concurrency.lockutils [req-5dfab5d7-4764-4f0d-bda1-224d1de78753 req-c66cb1d0-e79f-4479-aa53-29b955fdebc2 service nova] Releasing lock "refresh_cache-645abb7a-cf73-4fb0-a9a0-49205f060d6f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 727.073772] env[63531]: DEBUG oslo_concurrency.lockutils [None req-2f7bff43-205d-4df7-a4ce-48fd13394e84 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Lock "interface-f0276e34-1de3-482a-82c7-1439d40fd85e-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.108s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 727.085417] env[63531]: DEBUG oslo_concurrency.lockutils [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 727.159193] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1a515fd1-1aca-4043-ae4d-827540148b79 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Releasing lock "refresh_cache-0c948655-362d-4673-b3e6-3f28ec69ea3c" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 727.159193] env[63531]: DEBUG nova.compute.manager [None req-1a515fd1-1aca-4043-ae4d-827540148b79 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 727.159193] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-1a515fd1-1aca-4043-ae4d-827540148b79 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 727.159193] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d5b7f5f-3e6d-40e9-9c39-c4ac652c3796 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.167807] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a515fd1-1aca-4043-ae4d-827540148b79 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 727.168156] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9511e493-26cb-404f-95e3-c9ceec368c03 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.174696] env[63531]: DEBUG oslo_vmware.api [None req-1a515fd1-1aca-4043-ae4d-827540148b79 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Waiting for the task: (returnval){ [ 727.174696] env[63531]: value = "task-1117585" [ 727.174696] env[63531]: _type = "Task" [ 727.174696] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.183177] env[63531]: DEBUG oslo_vmware.api [None req-1a515fd1-1aca-4043-ae4d-827540148b79 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Task: {'id': task-1117585, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.250913] env[63531]: DEBUG nova.network.neutron [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: f39c8807-1e15-41cf-899b-a1fbe0695d58] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 727.287106] env[63531]: DEBUG nova.compute.manager [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 727.344764] env[63531]: DEBUG nova.network.neutron [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Successfully created port: f178c862-e753-40d4-b7cd-3eabe7eda39e {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 727.388641] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 727.411694] env[63531]: DEBUG nova.network.neutron [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: f39c8807-1e15-41cf-899b-a1fbe0695d58] Updating instance_info_cache with network_info: [{"id": "61792095-541d-4d6e-b33c-c35d459fe790", "address": "fa:16:3e:dc:d0:93", "network": {"id": "c3cf494e-23bd-480f-85d0-918fff65c5fa", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1373813036-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e2efa9eb41854d9cbcf96ebb8edbcfc9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4db2961d-273d-4634-9d06-a94fa9d384fb", "external-id": "nsx-vlan-transportzone-572", "segmentation_id": 572, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap61792095-54", "ovs_interfaceid": "61792095-541d-4d6e-b33c-c35d459fe790", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.497842] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52a4b365-0d22-556a-b6d9-283953a57539, 'name': SearchDatastore_Task, 'duration_secs': 0.0078} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.501167] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-07c67c1c-491c-4d64-bd79-dec1b333feab {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.503833] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Waiting for the task: (returnval){ [ 727.503833] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52198958-03da-ead0-684d-16023a802685" [ 727.503833] env[63531]: _type = "Task" [ 727.503833] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.514107] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52198958-03da-ead0-684d-16023a802685, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.671813] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b3e60e8-3d8c-4bfc-9e94-2fba7d740fe8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.686961] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d4020af-03fa-47e4-8580-3f132d6ee6c0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.690376] env[63531]: DEBUG oslo_vmware.api [None req-1a515fd1-1aca-4043-ae4d-827540148b79 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Task: {'id': task-1117585, 'name': PowerOffVM_Task, 'duration_secs': 0.124479} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.690534] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a515fd1-1aca-4043-ae4d-827540148b79 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 727.691029] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-1a515fd1-1aca-4043-ae4d-827540148b79 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 727.691302] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-078a7365-e7fc-4dc4-b83c-b95e4a91e542 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.722775] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9f17a7d-9dec-43ba-90e7-7afc918aa452 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.726622] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-1a515fd1-1aca-4043-ae4d-827540148b79 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 727.726889] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-1a515fd1-1aca-4043-ae4d-827540148b79 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 727.727093] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a515fd1-1aca-4043-ae4d-827540148b79 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Deleting the datastore file [datastore1] 0c948655-362d-4673-b3e6-3f28ec69ea3c {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 727.727589] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e4cb221b-ccc4-4b57-8417-089c2a54f29b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.736966] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4f1ad66-9624-48b6-9f2d-d87d58da8762 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.742029] env[63531]: DEBUG oslo_vmware.api [None req-1a515fd1-1aca-4043-ae4d-827540148b79 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Waiting for the task: (returnval){ [ 727.742029] env[63531]: value = "task-1117587" [ 727.742029] env[63531]: _type = "Task" [ 727.742029] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.764347] env[63531]: DEBUG nova.compute.provider_tree [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 727.768574] env[63531]: DEBUG nova.compute.manager [req-22892306-4789-4c6d-a364-1904647aa42b req-305b702e-b3d4-4e9c-8aab-f0af181e07e4 service nova] [instance: 7bc63fdd-4b50-486d-9681-ca0baa08f7b3] Received event network-vif-deleted-7348aaf4-66f3-47a2-9467-83aa99ae9fe4 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 727.774139] env[63531]: DEBUG oslo_vmware.api [None req-1a515fd1-1aca-4043-ae4d-827540148b79 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Task: {'id': task-1117587, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.914721] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Releasing lock "refresh_cache-f39c8807-1e15-41cf-899b-a1fbe0695d58" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 727.915100] env[63531]: DEBUG nova.compute.manager [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: f39c8807-1e15-41cf-899b-a1fbe0695d58] Instance network_info: |[{"id": "61792095-541d-4d6e-b33c-c35d459fe790", "address": "fa:16:3e:dc:d0:93", "network": {"id": "c3cf494e-23bd-480f-85d0-918fff65c5fa", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1373813036-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e2efa9eb41854d9cbcf96ebb8edbcfc9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4db2961d-273d-4634-9d06-a94fa9d384fb", "external-id": "nsx-vlan-transportzone-572", "segmentation_id": 572, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap61792095-54", "ovs_interfaceid": "61792095-541d-4d6e-b33c-c35d459fe790", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 727.915542] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: f39c8807-1e15-41cf-899b-a1fbe0695d58] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:dc:d0:93', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4db2961d-273d-4634-9d06-a94fa9d384fb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '61792095-541d-4d6e-b33c-c35d459fe790', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 727.924388] env[63531]: DEBUG oslo.service.loopingcall [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 727.924658] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f39c8807-1e15-41cf-899b-a1fbe0695d58] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 727.924903] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-11a105cc-e43c-4e33-8f06-6525b027d3d1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.944938] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 727.944938] env[63531]: value = "task-1117588" [ 727.944938] env[63531]: _type = "Task" [ 727.944938] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.953070] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117588, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.014311] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52198958-03da-ead0-684d-16023a802685, 'name': SearchDatastore_Task, 'duration_secs': 0.008889} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.014624] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 728.014900] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] 645abb7a-cf73-4fb0-a9a0-49205f060d6f/645abb7a-cf73-4fb0-a9a0-49205f060d6f.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 728.015192] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-33bb133a-dcf4-4614-bc8c-4191f9114e2b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.021667] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Waiting for the task: (returnval){ [ 728.021667] env[63531]: value = "task-1117589" [ 728.021667] env[63531]: _type = "Task" [ 728.021667] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.030886] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': task-1117589, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.187633] env[63531]: DEBUG nova.compute.manager [req-feb8da18-fe00-4c99-86bf-ae07694afbd5 req-c5d6a3dd-a190-4f98-8e20-4ef724cd744a service nova] [instance: f39c8807-1e15-41cf-899b-a1fbe0695d58] Received event network-vif-plugged-61792095-541d-4d6e-b33c-c35d459fe790 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 728.187830] env[63531]: DEBUG oslo_concurrency.lockutils [req-feb8da18-fe00-4c99-86bf-ae07694afbd5 req-c5d6a3dd-a190-4f98-8e20-4ef724cd744a service nova] Acquiring lock "f39c8807-1e15-41cf-899b-a1fbe0695d58-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 728.188092] env[63531]: DEBUG oslo_concurrency.lockutils [req-feb8da18-fe00-4c99-86bf-ae07694afbd5 req-c5d6a3dd-a190-4f98-8e20-4ef724cd744a service nova] Lock "f39c8807-1e15-41cf-899b-a1fbe0695d58-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 728.188200] env[63531]: DEBUG oslo_concurrency.lockutils [req-feb8da18-fe00-4c99-86bf-ae07694afbd5 req-c5d6a3dd-a190-4f98-8e20-4ef724cd744a service nova] Lock "f39c8807-1e15-41cf-899b-a1fbe0695d58-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 728.188370] env[63531]: DEBUG nova.compute.manager [req-feb8da18-fe00-4c99-86bf-ae07694afbd5 req-c5d6a3dd-a190-4f98-8e20-4ef724cd744a service nova] [instance: f39c8807-1e15-41cf-899b-a1fbe0695d58] No waiting events found dispatching network-vif-plugged-61792095-541d-4d6e-b33c-c35d459fe790 {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 728.188538] env[63531]: WARNING nova.compute.manager [req-feb8da18-fe00-4c99-86bf-ae07694afbd5 req-c5d6a3dd-a190-4f98-8e20-4ef724cd744a service nova] [instance: f39c8807-1e15-41cf-899b-a1fbe0695d58] Received unexpected event network-vif-plugged-61792095-541d-4d6e-b33c-c35d459fe790 for instance with vm_state building and task_state spawning. [ 728.188698] env[63531]: DEBUG nova.compute.manager [req-feb8da18-fe00-4c99-86bf-ae07694afbd5 req-c5d6a3dd-a190-4f98-8e20-4ef724cd744a service nova] [instance: f39c8807-1e15-41cf-899b-a1fbe0695d58] Received event network-changed-61792095-541d-4d6e-b33c-c35d459fe790 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 728.188853] env[63531]: DEBUG nova.compute.manager [req-feb8da18-fe00-4c99-86bf-ae07694afbd5 req-c5d6a3dd-a190-4f98-8e20-4ef724cd744a service nova] [instance: f39c8807-1e15-41cf-899b-a1fbe0695d58] Refreshing instance network info cache due to event network-changed-61792095-541d-4d6e-b33c-c35d459fe790. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 728.190009] env[63531]: DEBUG oslo_concurrency.lockutils [req-feb8da18-fe00-4c99-86bf-ae07694afbd5 req-c5d6a3dd-a190-4f98-8e20-4ef724cd744a service nova] Acquiring lock "refresh_cache-f39c8807-1e15-41cf-899b-a1fbe0695d58" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 728.190009] env[63531]: DEBUG oslo_concurrency.lockutils [req-feb8da18-fe00-4c99-86bf-ae07694afbd5 req-c5d6a3dd-a190-4f98-8e20-4ef724cd744a service nova] Acquired lock "refresh_cache-f39c8807-1e15-41cf-899b-a1fbe0695d58" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.190009] env[63531]: DEBUG nova.network.neutron [req-feb8da18-fe00-4c99-86bf-ae07694afbd5 req-c5d6a3dd-a190-4f98-8e20-4ef724cd744a service nova] [instance: f39c8807-1e15-41cf-899b-a1fbe0695d58] Refreshing network info cache for port 61792095-541d-4d6e-b33c-c35d459fe790 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 728.255141] env[63531]: DEBUG oslo_vmware.api [None req-1a515fd1-1aca-4043-ae4d-827540148b79 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Task: {'id': task-1117587, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.112229} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.255141] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a515fd1-1aca-4043-ae4d-827540148b79 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 728.255260] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-1a515fd1-1aca-4043-ae4d-827540148b79 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 728.255372] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-1a515fd1-1aca-4043-ae4d-827540148b79 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 728.255570] env[63531]: INFO nova.compute.manager [None req-1a515fd1-1aca-4043-ae4d-827540148b79 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Took 1.10 seconds to destroy the instance on the hypervisor. [ 728.255859] env[63531]: DEBUG oslo.service.loopingcall [None req-1a515fd1-1aca-4043-ae4d-827540148b79 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 728.256172] env[63531]: DEBUG nova.compute.manager [-] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 728.256291] env[63531]: DEBUG nova.network.neutron [-] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 728.268111] env[63531]: DEBUG nova.scheduler.client.report [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 728.273821] env[63531]: DEBUG nova.network.neutron [-] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 728.302931] env[63531]: DEBUG nova.compute.manager [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 728.332142] env[63531]: DEBUG nova.virt.hardware [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:17:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='86bdb6e9-50cf-4ad6-bad7-f8cb5dc21d1b',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-672645615',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 728.332401] env[63531]: DEBUG nova.virt.hardware [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 728.332588] env[63531]: DEBUG nova.virt.hardware [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 728.332863] env[63531]: DEBUG nova.virt.hardware [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 728.333044] env[63531]: DEBUG nova.virt.hardware [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 728.333233] env[63531]: DEBUG nova.virt.hardware [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 728.333532] env[63531]: DEBUG nova.virt.hardware [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 728.333756] env[63531]: DEBUG nova.virt.hardware [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 728.333989] env[63531]: DEBUG nova.virt.hardware [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 728.334251] env[63531]: DEBUG nova.virt.hardware [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 728.334469] env[63531]: DEBUG nova.virt.hardware [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 728.335522] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52ca30a3-21cf-4025-8504-20910e41f1ad {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.344953] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a4d64f3-279e-4e8d-8a57-c9400f902015 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.454709] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117588, 'name': CreateVM_Task, 'duration_secs': 0.384946} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.454871] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f39c8807-1e15-41cf-899b-a1fbe0695d58] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 728.455579] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 728.455739] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.456095] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 728.456359] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-16b84808-aabf-4c97-beb4-cb446890a5bd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.461712] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Waiting for the task: (returnval){ [ 728.461712] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52b76e6e-73d8-e953-303b-1b4817c47750" [ 728.461712] env[63531]: _type = "Task" [ 728.461712] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.473948] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52b76e6e-73d8-e953-303b-1b4817c47750, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.531634] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': task-1117589, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.482326} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.531910] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] 645abb7a-cf73-4fb0-a9a0-49205f060d6f/645abb7a-cf73-4fb0-a9a0-49205f060d6f.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 728.532143] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: 645abb7a-cf73-4fb0-a9a0-49205f060d6f] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 728.532392] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-66f77401-6a2c-46a4-a4bb-2737d979fb3f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.540043] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Waiting for the task: (returnval){ [ 728.540043] env[63531]: value = "task-1117590" [ 728.540043] env[63531]: _type = "Task" [ 728.540043] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.548580] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': task-1117590, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.774864] env[63531]: DEBUG oslo_concurrency.lockutils [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.499s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 728.775392] env[63531]: DEBUG nova.compute.manager [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 728.778900] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.011s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 728.780368] env[63531]: INFO nova.compute.claims [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 728.782876] env[63531]: DEBUG nova.network.neutron [-] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.980951] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52b76e6e-73d8-e953-303b-1b4817c47750, 'name': SearchDatastore_Task, 'duration_secs': 0.023912} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.984096] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 728.984479] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: f39c8807-1e15-41cf-899b-a1fbe0695d58] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 728.984866] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 728.985159] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.985455] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 728.985817] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c996a04e-e17a-4411-aa8a-55a093a4b5ea {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.994768] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 728.995065] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 728.996043] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-34c383e9-9041-4666-9068-c3c552acb734 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.007968] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Waiting for the task: (returnval){ [ 729.007968] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52a5b78e-c369-ebc0-ff30-0dc9c00bd9c4" [ 729.007968] env[63531]: _type = "Task" [ 729.007968] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.013474] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52a5b78e-c369-ebc0-ff30-0dc9c00bd9c4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.054811] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': task-1117590, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074717} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.054811] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: 645abb7a-cf73-4fb0-a9a0-49205f060d6f] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 729.054960] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7ff9ed9-b5b0-4089-930f-2e4438f73054 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.061678] env[63531]: DEBUG nova.network.neutron [req-feb8da18-fe00-4c99-86bf-ae07694afbd5 req-c5d6a3dd-a190-4f98-8e20-4ef724cd744a service nova] [instance: f39c8807-1e15-41cf-899b-a1fbe0695d58] Updated VIF entry in instance network info cache for port 61792095-541d-4d6e-b33c-c35d459fe790. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 729.061678] env[63531]: DEBUG nova.network.neutron [req-feb8da18-fe00-4c99-86bf-ae07694afbd5 req-c5d6a3dd-a190-4f98-8e20-4ef724cd744a service nova] [instance: f39c8807-1e15-41cf-899b-a1fbe0695d58] Updating instance_info_cache with network_info: [{"id": "61792095-541d-4d6e-b33c-c35d459fe790", "address": "fa:16:3e:dc:d0:93", "network": {"id": "c3cf494e-23bd-480f-85d0-918fff65c5fa", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1373813036-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e2efa9eb41854d9cbcf96ebb8edbcfc9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4db2961d-273d-4634-9d06-a94fa9d384fb", "external-id": "nsx-vlan-transportzone-572", "segmentation_id": 572, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap61792095-54", "ovs_interfaceid": "61792095-541d-4d6e-b33c-c35d459fe790", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 729.083884] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: 645abb7a-cf73-4fb0-a9a0-49205f060d6f] Reconfiguring VM instance instance-00000025 to attach disk [datastore2] 645abb7a-cf73-4fb0-a9a0-49205f060d6f/645abb7a-cf73-4fb0-a9a0-49205f060d6f.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 729.083884] env[63531]: DEBUG oslo_concurrency.lockutils [req-feb8da18-fe00-4c99-86bf-ae07694afbd5 req-c5d6a3dd-a190-4f98-8e20-4ef724cd744a service nova] Releasing lock "refresh_cache-f39c8807-1e15-41cf-899b-a1fbe0695d58" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 729.083884] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-05d3e12b-8921-4ce1-8cd7-71eccee94188 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.103466] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Waiting for the task: (returnval){ [ 729.103466] env[63531]: value = "task-1117591" [ 729.103466] env[63531]: _type = "Task" [ 729.103466] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.113600] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': task-1117591, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.212270] env[63531]: DEBUG oslo_concurrency.lockutils [None req-524a4683-a8e1-4d4b-87d9-f141433baa74 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Acquiring lock "f0276e34-1de3-482a-82c7-1439d40fd85e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 729.212270] env[63531]: DEBUG oslo_concurrency.lockutils [None req-524a4683-a8e1-4d4b-87d9-f141433baa74 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Lock "f0276e34-1de3-482a-82c7-1439d40fd85e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 729.212270] env[63531]: DEBUG oslo_concurrency.lockutils [None req-524a4683-a8e1-4d4b-87d9-f141433baa74 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Acquiring lock "f0276e34-1de3-482a-82c7-1439d40fd85e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 729.212270] env[63531]: DEBUG oslo_concurrency.lockutils [None req-524a4683-a8e1-4d4b-87d9-f141433baa74 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Lock "f0276e34-1de3-482a-82c7-1439d40fd85e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 729.212678] env[63531]: DEBUG oslo_concurrency.lockutils [None req-524a4683-a8e1-4d4b-87d9-f141433baa74 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Lock "f0276e34-1de3-482a-82c7-1439d40fd85e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 729.213399] env[63531]: INFO nova.compute.manager [None req-524a4683-a8e1-4d4b-87d9-f141433baa74 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] [instance: f0276e34-1de3-482a-82c7-1439d40fd85e] Terminating instance [ 729.215350] env[63531]: DEBUG nova.compute.manager [None req-524a4683-a8e1-4d4b-87d9-f141433baa74 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] [instance: f0276e34-1de3-482a-82c7-1439d40fd85e] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 729.215660] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-524a4683-a8e1-4d4b-87d9-f141433baa74 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] [instance: f0276e34-1de3-482a-82c7-1439d40fd85e] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 729.216572] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90a30c1a-3354-4829-af9f-a3aa9334aeb4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.226039] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-524a4683-a8e1-4d4b-87d9-f141433baa74 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] [instance: f0276e34-1de3-482a-82c7-1439d40fd85e] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 729.226505] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2939ad6b-842c-4193-84da-cceafd9bc81e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.233233] env[63531]: DEBUG oslo_vmware.api [None req-524a4683-a8e1-4d4b-87d9-f141433baa74 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Waiting for the task: (returnval){ [ 729.233233] env[63531]: value = "task-1117592" [ 729.233233] env[63531]: _type = "Task" [ 729.233233] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.241425] env[63531]: DEBUG nova.network.neutron [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Successfully updated port: f178c862-e753-40d4-b7cd-3eabe7eda39e {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 729.245135] env[63531]: DEBUG oslo_vmware.api [None req-524a4683-a8e1-4d4b-87d9-f141433baa74 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Task: {'id': task-1117592, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.281415] env[63531]: DEBUG nova.compute.utils [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 729.282753] env[63531]: DEBUG nova.compute.manager [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 729.282847] env[63531]: DEBUG nova.network.neutron [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 729.287980] env[63531]: INFO nova.compute.manager [-] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Took 1.03 seconds to deallocate network for instance. [ 729.371029] env[63531]: DEBUG nova.policy [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b148fcad75cb4abfabd4d00d392f4136', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a634a61020b94d62a8fb37a3f96acc10', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 729.516314] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52a5b78e-c369-ebc0-ff30-0dc9c00bd9c4, 'name': SearchDatastore_Task, 'duration_secs': 0.010357} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.517274] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-30521026-c1c8-406e-b3e2-d8d58c83e7f2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.522371] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Waiting for the task: (returnval){ [ 729.522371] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52e70114-11b6-3236-d54e-01841335e163" [ 729.522371] env[63531]: _type = "Task" [ 729.522371] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.530041] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52e70114-11b6-3236-d54e-01841335e163, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.615405] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': task-1117591, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.743368] env[63531]: DEBUG oslo_vmware.api [None req-524a4683-a8e1-4d4b-87d9-f141433baa74 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Task: {'id': task-1117592, 'name': PowerOffVM_Task, 'duration_secs': 0.21147} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.743649] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-524a4683-a8e1-4d4b-87d9-f141433baa74 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] [instance: f0276e34-1de3-482a-82c7-1439d40fd85e] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 729.743819] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-524a4683-a8e1-4d4b-87d9-f141433baa74 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] [instance: f0276e34-1de3-482a-82c7-1439d40fd85e] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 729.744083] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-af52a577-bf2e-487a-af4a-a84fdeda59bd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.748377] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Acquiring lock "refresh_cache-5e9042b3-4dc2-4fa3-a664-c4b49a22e400" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 729.748377] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Acquired lock "refresh_cache-5e9042b3-4dc2-4fa3-a664-c4b49a22e400" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.748377] env[63531]: DEBUG nova.network.neutron [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 729.790275] env[63531]: DEBUG nova.compute.manager [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 729.797168] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1a515fd1-1aca-4043-ae4d-827540148b79 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 729.812525] env[63531]: DEBUG nova.network.neutron [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Successfully created port: 91733d5a-84d6-48df-ba4d-7e362a642282 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 729.821386] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-524a4683-a8e1-4d4b-87d9-f141433baa74 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] [instance: f0276e34-1de3-482a-82c7-1439d40fd85e] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 729.821493] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-524a4683-a8e1-4d4b-87d9-f141433baa74 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] [instance: f0276e34-1de3-482a-82c7-1439d40fd85e] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 729.821855] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-524a4683-a8e1-4d4b-87d9-f141433baa74 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Deleting the datastore file [datastore1] f0276e34-1de3-482a-82c7-1439d40fd85e {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 729.825137] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5f1719e2-0b80-4e69-8637-fc0d2a0e6985 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.834930] env[63531]: DEBUG nova.compute.manager [req-d6039c7f-4120-4b1e-b47b-a58bcd22f11f req-9d2a17fd-d9bb-4df8-aa59-fdb242b2d6cf service nova] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Received event network-vif-plugged-f178c862-e753-40d4-b7cd-3eabe7eda39e {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 729.835192] env[63531]: DEBUG oslo_concurrency.lockutils [req-d6039c7f-4120-4b1e-b47b-a58bcd22f11f req-9d2a17fd-d9bb-4df8-aa59-fdb242b2d6cf service nova] Acquiring lock "5e9042b3-4dc2-4fa3-a664-c4b49a22e400-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 729.835499] env[63531]: DEBUG oslo_concurrency.lockutils [req-d6039c7f-4120-4b1e-b47b-a58bcd22f11f req-9d2a17fd-d9bb-4df8-aa59-fdb242b2d6cf service nova] Lock "5e9042b3-4dc2-4fa3-a664-c4b49a22e400-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 729.835629] env[63531]: DEBUG oslo_concurrency.lockutils [req-d6039c7f-4120-4b1e-b47b-a58bcd22f11f req-9d2a17fd-d9bb-4df8-aa59-fdb242b2d6cf service nova] Lock "5e9042b3-4dc2-4fa3-a664-c4b49a22e400-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 729.835741] env[63531]: DEBUG nova.compute.manager [req-d6039c7f-4120-4b1e-b47b-a58bcd22f11f req-9d2a17fd-d9bb-4df8-aa59-fdb242b2d6cf service nova] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] No waiting events found dispatching network-vif-plugged-f178c862-e753-40d4-b7cd-3eabe7eda39e {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 729.835903] env[63531]: WARNING nova.compute.manager [req-d6039c7f-4120-4b1e-b47b-a58bcd22f11f req-9d2a17fd-d9bb-4df8-aa59-fdb242b2d6cf service nova] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Received unexpected event network-vif-plugged-f178c862-e753-40d4-b7cd-3eabe7eda39e for instance with vm_state building and task_state spawning. [ 729.836065] env[63531]: DEBUG nova.compute.manager [req-d6039c7f-4120-4b1e-b47b-a58bcd22f11f req-9d2a17fd-d9bb-4df8-aa59-fdb242b2d6cf service nova] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Received event network-changed-f178c862-e753-40d4-b7cd-3eabe7eda39e {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 729.836269] env[63531]: DEBUG nova.compute.manager [req-d6039c7f-4120-4b1e-b47b-a58bcd22f11f req-9d2a17fd-d9bb-4df8-aa59-fdb242b2d6cf service nova] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Refreshing instance network info cache due to event network-changed-f178c862-e753-40d4-b7cd-3eabe7eda39e. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 729.836700] env[63531]: DEBUG oslo_concurrency.lockutils [req-d6039c7f-4120-4b1e-b47b-a58bcd22f11f req-9d2a17fd-d9bb-4df8-aa59-fdb242b2d6cf service nova] Acquiring lock "refresh_cache-5e9042b3-4dc2-4fa3-a664-c4b49a22e400" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 729.838698] env[63531]: DEBUG oslo_vmware.api [None req-524a4683-a8e1-4d4b-87d9-f141433baa74 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Waiting for the task: (returnval){ [ 729.838698] env[63531]: value = "task-1117594" [ 729.838698] env[63531]: _type = "Task" [ 729.838698] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.850948] env[63531]: DEBUG oslo_vmware.api [None req-524a4683-a8e1-4d4b-87d9-f141433baa74 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Task: {'id': task-1117594, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.037891] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52e70114-11b6-3236-d54e-01841335e163, 'name': SearchDatastore_Task, 'duration_secs': 0.047881} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.038185] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 730.038442] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] f39c8807-1e15-41cf-899b-a1fbe0695d58/f39c8807-1e15-41cf-899b-a1fbe0695d58.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 730.038782] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4e75a0ba-4554-4428-ade1-0ba8aca8af3c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.051507] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Waiting for the task: (returnval){ [ 730.051507] env[63531]: value = "task-1117595" [ 730.051507] env[63531]: _type = "Task" [ 730.051507] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.062492] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': task-1117595, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.116270] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': task-1117591, 'name': ReconfigVM_Task, 'duration_secs': 0.575386} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.116548] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: 645abb7a-cf73-4fb0-a9a0-49205f060d6f] Reconfigured VM instance instance-00000025 to attach disk [datastore2] 645abb7a-cf73-4fb0-a9a0-49205f060d6f/645abb7a-cf73-4fb0-a9a0-49205f060d6f.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 730.117209] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-df6b86f7-ca77-4210-906e-a1a9d6d8ffe5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.124039] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Waiting for the task: (returnval){ [ 730.124039] env[63531]: value = "task-1117596" [ 730.124039] env[63531]: _type = "Task" [ 730.124039] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.132886] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': task-1117596, 'name': Rename_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.174261] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0130981-1cb2-45a8-ae6a-e4d8bca1a3f1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.181860] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f16980d5-e288-4a4a-b540-f2335814d7d5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.216555] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-effe5fc4-44a8-4167-a273-8136206d3c26 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.224391] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-773f9581-5c7b-45b6-b128-22194ef22540 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.238624] env[63531]: DEBUG nova.compute.provider_tree [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 730.288854] env[63531]: DEBUG nova.network.neutron [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 730.352365] env[63531]: DEBUG oslo_vmware.api [None req-524a4683-a8e1-4d4b-87d9-f141433baa74 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Task: {'id': task-1117594, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.479397} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.352628] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-524a4683-a8e1-4d4b-87d9-f141433baa74 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 730.352810] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-524a4683-a8e1-4d4b-87d9-f141433baa74 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] [instance: f0276e34-1de3-482a-82c7-1439d40fd85e] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 730.352988] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-524a4683-a8e1-4d4b-87d9-f141433baa74 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] [instance: f0276e34-1de3-482a-82c7-1439d40fd85e] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 730.353179] env[63531]: INFO nova.compute.manager [None req-524a4683-a8e1-4d4b-87d9-f141433baa74 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] [instance: f0276e34-1de3-482a-82c7-1439d40fd85e] Took 1.14 seconds to destroy the instance on the hypervisor. [ 730.353427] env[63531]: DEBUG oslo.service.loopingcall [None req-524a4683-a8e1-4d4b-87d9-f141433baa74 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 730.353609] env[63531]: DEBUG nova.compute.manager [-] [instance: f0276e34-1de3-482a-82c7-1439d40fd85e] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 730.353703] env[63531]: DEBUG nova.network.neutron [-] [instance: f0276e34-1de3-482a-82c7-1439d40fd85e] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 730.475931] env[63531]: DEBUG nova.network.neutron [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Updating instance_info_cache with network_info: [{"id": "f178c862-e753-40d4-b7cd-3eabe7eda39e", "address": "fa:16:3e:a6:d3:d2", "network": {"id": "5439232a-ec6d-4184-a9f1-f7be5c1526bc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.35", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "54fbef8a6ee0488f820706d2036e9c99", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f267bcdd-0daa-4337-9709-5fc060c267d8", "external-id": "nsx-vlan-transportzone-308", "segmentation_id": 308, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf178c862-e7", "ovs_interfaceid": "f178c862-e753-40d4-b7cd-3eabe7eda39e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 730.564716] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': task-1117595, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.641079] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': task-1117596, 'name': Rename_Task, 'duration_secs': 0.235236} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.641504] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: 645abb7a-cf73-4fb0-a9a0-49205f060d6f] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 730.641826] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8933d04c-e19e-46be-9195-637cb821a084 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.651608] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Waiting for the task: (returnval){ [ 730.651608] env[63531]: value = "task-1117597" [ 730.651608] env[63531]: _type = "Task" [ 730.651608] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.662251] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': task-1117597, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.744659] env[63531]: DEBUG nova.scheduler.client.report [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 730.804682] env[63531]: DEBUG nova.compute.manager [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 730.838172] env[63531]: DEBUG nova.virt.hardware [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 730.838405] env[63531]: DEBUG nova.virt.hardware [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 730.838554] env[63531]: DEBUG nova.virt.hardware [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 730.838728] env[63531]: DEBUG nova.virt.hardware [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 730.838866] env[63531]: DEBUG nova.virt.hardware [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 730.839028] env[63531]: DEBUG nova.virt.hardware [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 730.839271] env[63531]: DEBUG nova.virt.hardware [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 730.839422] env[63531]: DEBUG nova.virt.hardware [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 730.841990] env[63531]: DEBUG nova.virt.hardware [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 730.841990] env[63531]: DEBUG nova.virt.hardware [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 730.841990] env[63531]: DEBUG nova.virt.hardware [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 730.841990] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-954e0680-12b2-4c63-99e6-57212310854d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.850247] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8aa776b0-ebf1-4900-ab6f-378b27f63c62 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.981539] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Releasing lock "refresh_cache-5e9042b3-4dc2-4fa3-a664-c4b49a22e400" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 730.981932] env[63531]: DEBUG nova.compute.manager [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Instance network_info: |[{"id": "f178c862-e753-40d4-b7cd-3eabe7eda39e", "address": "fa:16:3e:a6:d3:d2", "network": {"id": "5439232a-ec6d-4184-a9f1-f7be5c1526bc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.35", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "54fbef8a6ee0488f820706d2036e9c99", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f267bcdd-0daa-4337-9709-5fc060c267d8", "external-id": "nsx-vlan-transportzone-308", "segmentation_id": 308, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf178c862-e7", "ovs_interfaceid": "f178c862-e753-40d4-b7cd-3eabe7eda39e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 730.982339] env[63531]: DEBUG oslo_concurrency.lockutils [req-d6039c7f-4120-4b1e-b47b-a58bcd22f11f req-9d2a17fd-d9bb-4df8-aa59-fdb242b2d6cf service nova] Acquired lock "refresh_cache-5e9042b3-4dc2-4fa3-a664-c4b49a22e400" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.982557] env[63531]: DEBUG nova.network.neutron [req-d6039c7f-4120-4b1e-b47b-a58bcd22f11f req-9d2a17fd-d9bb-4df8-aa59-fdb242b2d6cf service nova] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Refreshing network info cache for port f178c862-e753-40d4-b7cd-3eabe7eda39e {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 730.983977] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a6:d3:d2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f267bcdd-0daa-4337-9709-5fc060c267d8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f178c862-e753-40d4-b7cd-3eabe7eda39e', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 730.991861] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Creating folder: Project (5f0d10f679de4fa4bd9d725951e536c1). Parent ref: group-v244585. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 730.993067] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-01e72cd8-a48d-4833-8eb1-d6d60553c909 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.004269] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Created folder: Project (5f0d10f679de4fa4bd9d725951e536c1) in parent group-v244585. [ 731.004469] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Creating folder: Instances. Parent ref: group-v244620. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 731.004917] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5c000e3c-0908-482a-b38e-064c93e665ee {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.014881] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Created folder: Instances in parent group-v244620. [ 731.015127] env[63531]: DEBUG oslo.service.loopingcall [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 731.015325] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 731.015531] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9a146a0a-b74f-4315-9124-34046a48789f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.036837] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 731.036837] env[63531]: value = "task-1117600" [ 731.036837] env[63531]: _type = "Task" [ 731.036837] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.044491] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117600, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.062599] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': task-1117595, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.67358} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.062857] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] f39c8807-1e15-41cf-899b-a1fbe0695d58/f39c8807-1e15-41cf-899b-a1fbe0695d58.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 731.063087] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: f39c8807-1e15-41cf-899b-a1fbe0695d58] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 731.063335] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-429b0006-8f75-4ae1-8265-ea696489906b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.072480] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Waiting for the task: (returnval){ [ 731.072480] env[63531]: value = "task-1117601" [ 731.072480] env[63531]: _type = "Task" [ 731.072480] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.082063] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': task-1117601, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.161858] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': task-1117597, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.250136] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.471s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 731.250745] env[63531]: DEBUG nova.compute.manager [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 731.254341] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.435s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 731.255620] env[63531]: INFO nova.compute.claims [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 731.313518] env[63531]: DEBUG nova.network.neutron [-] [instance: f0276e34-1de3-482a-82c7-1439d40fd85e] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.551173] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117600, 'name': CreateVM_Task} progress is 25%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.581876] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': task-1117601, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.210143} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.582170] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: f39c8807-1e15-41cf-899b-a1fbe0695d58] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 731.583088] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-779bfc9f-0a3c-421a-b4f1-f5883aea405a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.608414] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: f39c8807-1e15-41cf-899b-a1fbe0695d58] Reconfiguring VM instance instance-00000026 to attach disk [datastore1] f39c8807-1e15-41cf-899b-a1fbe0695d58/f39c8807-1e15-41cf-899b-a1fbe0695d58.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 731.608737] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6f565782-129a-4ced-b447-4d9936fde72a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.627752] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Waiting for the task: (returnval){ [ 731.627752] env[63531]: value = "task-1117602" [ 731.627752] env[63531]: _type = "Task" [ 731.627752] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.635789] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': task-1117602, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.660938] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': task-1117597, 'name': PowerOnVM_Task, 'duration_secs': 0.934835} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.661246] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: 645abb7a-cf73-4fb0-a9a0-49205f060d6f] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 731.661454] env[63531]: INFO nova.compute.manager [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: 645abb7a-cf73-4fb0-a9a0-49205f060d6f] Took 8.39 seconds to spawn the instance on the hypervisor. [ 731.661632] env[63531]: DEBUG nova.compute.manager [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: 645abb7a-cf73-4fb0-a9a0-49205f060d6f] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 731.662437] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42c5a30d-f30d-4f0c-bb0e-6df4ee0111d7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.756386] env[63531]: DEBUG nova.compute.utils [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 731.758492] env[63531]: DEBUG nova.compute.manager [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 731.758677] env[63531]: DEBUG nova.network.neutron [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 731.802739] env[63531]: DEBUG nova.policy [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6d9c6a44f8944b70b498d67a89e7ee2c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fed6b4862d76469b87ecf16dc6dce177', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 731.816500] env[63531]: INFO nova.compute.manager [-] [instance: f0276e34-1de3-482a-82c7-1439d40fd85e] Took 1.46 seconds to deallocate network for instance. [ 731.899100] env[63531]: DEBUG nova.compute.manager [req-ad7c4f4b-1dd8-4cc2-b7bb-e02d4dc7abd9 req-78d6db1f-31c4-4dad-919d-a37e6ab1efb3 service nova] [instance: f0276e34-1de3-482a-82c7-1439d40fd85e] Received event network-vif-deleted-08a2f25f-31d7-4af6-9116-f946e065155a {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 731.899575] env[63531]: DEBUG nova.compute.manager [req-ad7c4f4b-1dd8-4cc2-b7bb-e02d4dc7abd9 req-78d6db1f-31c4-4dad-919d-a37e6ab1efb3 service nova] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Received event network-vif-plugged-91733d5a-84d6-48df-ba4d-7e362a642282 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 731.899839] env[63531]: DEBUG oslo_concurrency.lockutils [req-ad7c4f4b-1dd8-4cc2-b7bb-e02d4dc7abd9 req-78d6db1f-31c4-4dad-919d-a37e6ab1efb3 service nova] Acquiring lock "c3920e60-b27f-496b-b616-afb23ab0bc26-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 731.900489] env[63531]: DEBUG oslo_concurrency.lockutils [req-ad7c4f4b-1dd8-4cc2-b7bb-e02d4dc7abd9 req-78d6db1f-31c4-4dad-919d-a37e6ab1efb3 service nova] Lock "c3920e60-b27f-496b-b616-afb23ab0bc26-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 731.900881] env[63531]: DEBUG oslo_concurrency.lockutils [req-ad7c4f4b-1dd8-4cc2-b7bb-e02d4dc7abd9 req-78d6db1f-31c4-4dad-919d-a37e6ab1efb3 service nova] Lock "c3920e60-b27f-496b-b616-afb23ab0bc26-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 731.901303] env[63531]: DEBUG nova.compute.manager [req-ad7c4f4b-1dd8-4cc2-b7bb-e02d4dc7abd9 req-78d6db1f-31c4-4dad-919d-a37e6ab1efb3 service nova] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] No waiting events found dispatching network-vif-plugged-91733d5a-84d6-48df-ba4d-7e362a642282 {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 731.901521] env[63531]: WARNING nova.compute.manager [req-ad7c4f4b-1dd8-4cc2-b7bb-e02d4dc7abd9 req-78d6db1f-31c4-4dad-919d-a37e6ab1efb3 service nova] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Received unexpected event network-vif-plugged-91733d5a-84d6-48df-ba4d-7e362a642282 for instance with vm_state building and task_state spawning. [ 731.927175] env[63531]: DEBUG nova.network.neutron [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Successfully updated port: 91733d5a-84d6-48df-ba4d-7e362a642282 {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 732.049627] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117600, 'name': CreateVM_Task, 'duration_secs': 0.709207} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.049816] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 732.050713] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 732.050713] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.051079] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 732.051217] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5ea2b49a-683d-48d0-a86b-43cbe92049e4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.055875] env[63531]: DEBUG oslo_vmware.api [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Waiting for the task: (returnval){ [ 732.055875] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52fce7ed-1f2c-0fcc-7d0e-a21ed89b88b9" [ 732.055875] env[63531]: _type = "Task" [ 732.055875] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.063660] env[63531]: DEBUG oslo_vmware.api [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52fce7ed-1f2c-0fcc-7d0e-a21ed89b88b9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.110943] env[63531]: DEBUG nova.network.neutron [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Successfully created port: c173088d-cfd8-4f05-8936-1d0d5927ae34 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 732.137979] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': task-1117602, 'name': ReconfigVM_Task, 'duration_secs': 0.301704} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.138324] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: f39c8807-1e15-41cf-899b-a1fbe0695d58] Reconfigured VM instance instance-00000026 to attach disk [datastore1] f39c8807-1e15-41cf-899b-a1fbe0695d58/f39c8807-1e15-41cf-899b-a1fbe0695d58.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 732.138942] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-92eb8f2e-8e68-47f7-bc30-f457583f4c14 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.145425] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Waiting for the task: (returnval){ [ 732.145425] env[63531]: value = "task-1117603" [ 732.145425] env[63531]: _type = "Task" [ 732.145425] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.153089] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': task-1117603, 'name': Rename_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.159684] env[63531]: DEBUG nova.network.neutron [req-d6039c7f-4120-4b1e-b47b-a58bcd22f11f req-9d2a17fd-d9bb-4df8-aa59-fdb242b2d6cf service nova] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Updated VIF entry in instance network info cache for port f178c862-e753-40d4-b7cd-3eabe7eda39e. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 732.160104] env[63531]: DEBUG nova.network.neutron [req-d6039c7f-4120-4b1e-b47b-a58bcd22f11f req-9d2a17fd-d9bb-4df8-aa59-fdb242b2d6cf service nova] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Updating instance_info_cache with network_info: [{"id": "f178c862-e753-40d4-b7cd-3eabe7eda39e", "address": "fa:16:3e:a6:d3:d2", "network": {"id": "5439232a-ec6d-4184-a9f1-f7be5c1526bc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.35", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "54fbef8a6ee0488f820706d2036e9c99", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f267bcdd-0daa-4337-9709-5fc060c267d8", "external-id": "nsx-vlan-transportzone-308", "segmentation_id": 308, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf178c862-e7", "ovs_interfaceid": "f178c862-e753-40d4-b7cd-3eabe7eda39e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 732.181426] env[63531]: INFO nova.compute.manager [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: 645abb7a-cf73-4fb0-a9a0-49205f060d6f] Took 43.42 seconds to build instance. [ 732.262172] env[63531]: DEBUG nova.compute.manager [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 732.326625] env[63531]: DEBUG oslo_concurrency.lockutils [None req-524a4683-a8e1-4d4b-87d9-f141433baa74 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 732.433015] env[63531]: DEBUG oslo_concurrency.lockutils [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Acquiring lock "refresh_cache-c3920e60-b27f-496b-b616-afb23ab0bc26" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 732.433206] env[63531]: DEBUG oslo_concurrency.lockutils [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Acquired lock "refresh_cache-c3920e60-b27f-496b-b616-afb23ab0bc26" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.433323] env[63531]: DEBUG nova.network.neutron [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 732.568184] env[63531]: DEBUG oslo_vmware.api [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52fce7ed-1f2c-0fcc-7d0e-a21ed89b88b9, 'name': SearchDatastore_Task, 'duration_secs': 0.039523} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.568538] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 732.568703] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 732.569075] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 732.569125] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.569361] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 732.569528] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a478e9d3-2e3d-427d-a6de-fdd9a361c1b8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.577317] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 732.577490] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 732.580682] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8f6954fd-be82-4c97-96b8-054fc79f0e42 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.585825] env[63531]: DEBUG oslo_vmware.api [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Waiting for the task: (returnval){ [ 732.585825] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52f80551-a97a-e592-4ee6-ba7ad7fde9d7" [ 732.585825] env[63531]: _type = "Task" [ 732.585825] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.594363] env[63531]: DEBUG oslo_vmware.api [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52f80551-a97a-e592-4ee6-ba7ad7fde9d7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.619226] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-395be51e-3241-40a3-8322-2f0e98469fe3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.626488] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dbc619b-06bc-4d44-9ca7-c331b340bd6b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.660961] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6ce003a-18ef-46e7-ae15-cef41901d812 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.664098] env[63531]: DEBUG oslo_concurrency.lockutils [req-d6039c7f-4120-4b1e-b47b-a58bcd22f11f req-9d2a17fd-d9bb-4df8-aa59-fdb242b2d6cf service nova] Releasing lock "refresh_cache-5e9042b3-4dc2-4fa3-a664-c4b49a22e400" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 732.670972] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-167e370f-eaf0-47fa-be77-e737c9c8ee72 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.674623] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': task-1117603, 'name': Rename_Task, 'duration_secs': 0.148439} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.674850] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: f39c8807-1e15-41cf-899b-a1fbe0695d58] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 732.675379] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cf9ab676-e782-4ed2-bce3-d993184c7b90 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.684876] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Lock "645abb7a-cf73-4fb0-a9a0-49205f060d6f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 126.544s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 732.685330] env[63531]: DEBUG nova.compute.provider_tree [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 732.689060] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Waiting for the task: (returnval){ [ 732.689060] env[63531]: value = "task-1117604" [ 732.689060] env[63531]: _type = "Task" [ 732.689060] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.698434] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': task-1117604, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.768737] env[63531]: INFO nova.virt.block_device [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Booting with volume 2a24c282-f681-40bf-a19c-5c711d342687 at /dev/sda [ 732.805409] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cbcee21d-26dd-478b-af41-b36979735f86 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.814676] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1741b42-94a6-4e3c-989d-3b1a0c20cab2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.839039] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9e5edbbd-d870-4d29-9db1-d190caf902d6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.847084] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df6e55bd-9856-44c3-90fe-4493e24d9d9b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.872050] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-819bcc6e-0298-4574-a56b-f2a7e1c347d4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.878796] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21196e98-b7e0-415f-8b79-13ddd0b08ce3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.894845] env[63531]: DEBUG nova.virt.block_device [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Updating existing volume attachment record: 95f063b2-3b3b-458f-a9f8-d98f48414b1a {{(pid=63531) _volume_attach /opt/stack/nova/nova/virt/block_device.py:679}} [ 732.984031] env[63531]: DEBUG nova.network.neutron [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 733.097316] env[63531]: DEBUG oslo_vmware.api [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52f80551-a97a-e592-4ee6-ba7ad7fde9d7, 'name': SearchDatastore_Task, 'duration_secs': 0.035185} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.100353] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-576bffa2-6e59-4b84-a884-bda7b53821bb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.105961] env[63531]: DEBUG oslo_vmware.api [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Waiting for the task: (returnval){ [ 733.105961] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52dc50db-94c1-d481-fd6d-fc64634b33c4" [ 733.105961] env[63531]: _type = "Task" [ 733.105961] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.115024] env[63531]: DEBUG oslo_vmware.api [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52dc50db-94c1-d481-fd6d-fc64634b33c4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.138225] env[63531]: DEBUG nova.network.neutron [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Updating instance_info_cache with network_info: [{"id": "91733d5a-84d6-48df-ba4d-7e362a642282", "address": "fa:16:3e:df:a1:31", "network": {"id": "d28b2087-329c-4917-aa6e-83ab83a63a74", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1054369059-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a634a61020b94d62a8fb37a3f96acc10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91733d5a-84", "ovs_interfaceid": "91733d5a-84d6-48df-ba4d-7e362a642282", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.194016] env[63531]: DEBUG nova.scheduler.client.report [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 733.197722] env[63531]: DEBUG nova.compute.manager [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 733.212994] env[63531]: DEBUG oslo_vmware.api [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': task-1117604, 'name': PowerOnVM_Task, 'duration_secs': 0.442009} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.213352] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: f39c8807-1e15-41cf-899b-a1fbe0695d58] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 733.213566] env[63531]: INFO nova.compute.manager [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: f39c8807-1e15-41cf-899b-a1fbe0695d58] Took 7.46 seconds to spawn the instance on the hypervisor. [ 733.213768] env[63531]: DEBUG nova.compute.manager [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: f39c8807-1e15-41cf-899b-a1fbe0695d58] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 733.214567] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ad16179-62f7-4128-b637-95cbdf0d3b94 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.589705] env[63531]: DEBUG nova.compute.manager [req-170d7b41-b6b8-44d9-907b-a7d2c4c3a787 req-a282f1ba-e5a7-4c45-b4ee-970dad3185e7 service nova] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Received event network-vif-plugged-c173088d-cfd8-4f05-8936-1d0d5927ae34 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 733.589951] env[63531]: DEBUG oslo_concurrency.lockutils [req-170d7b41-b6b8-44d9-907b-a7d2c4c3a787 req-a282f1ba-e5a7-4c45-b4ee-970dad3185e7 service nova] Acquiring lock "410ddf7a-e45b-4769-a3db-5363ce2096a7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.591197] env[63531]: DEBUG oslo_concurrency.lockutils [req-170d7b41-b6b8-44d9-907b-a7d2c4c3a787 req-a282f1ba-e5a7-4c45-b4ee-970dad3185e7 service nova] Lock "410ddf7a-e45b-4769-a3db-5363ce2096a7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.591197] env[63531]: DEBUG oslo_concurrency.lockutils [req-170d7b41-b6b8-44d9-907b-a7d2c4c3a787 req-a282f1ba-e5a7-4c45-b4ee-970dad3185e7 service nova] Lock "410ddf7a-e45b-4769-a3db-5363ce2096a7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 733.591197] env[63531]: DEBUG nova.compute.manager [req-170d7b41-b6b8-44d9-907b-a7d2c4c3a787 req-a282f1ba-e5a7-4c45-b4ee-970dad3185e7 service nova] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] No waiting events found dispatching network-vif-plugged-c173088d-cfd8-4f05-8936-1d0d5927ae34 {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 733.591756] env[63531]: WARNING nova.compute.manager [req-170d7b41-b6b8-44d9-907b-a7d2c4c3a787 req-a282f1ba-e5a7-4c45-b4ee-970dad3185e7 service nova] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Received unexpected event network-vif-plugged-c173088d-cfd8-4f05-8936-1d0d5927ae34 for instance with vm_state building and task_state block_device_mapping. [ 733.617958] env[63531]: DEBUG oslo_vmware.api [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52dc50db-94c1-d481-fd6d-fc64634b33c4, 'name': SearchDatastore_Task, 'duration_secs': 0.025025} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.618377] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.618663] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 5e9042b3-4dc2-4fa3-a664-c4b49a22e400/5e9042b3-4dc2-4fa3-a664-c4b49a22e400.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 733.618997] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e5ffcc60-de61-4c8e-a0b9-7e2a77d8b1bd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.626844] env[63531]: DEBUG oslo_vmware.api [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Waiting for the task: (returnval){ [ 733.626844] env[63531]: value = "task-1117605" [ 733.626844] env[63531]: _type = "Task" [ 733.626844] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.640653] env[63531]: DEBUG oslo_concurrency.lockutils [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Releasing lock "refresh_cache-c3920e60-b27f-496b-b616-afb23ab0bc26" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.640970] env[63531]: DEBUG nova.compute.manager [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Instance network_info: |[{"id": "91733d5a-84d6-48df-ba4d-7e362a642282", "address": "fa:16:3e:df:a1:31", "network": {"id": "d28b2087-329c-4917-aa6e-83ab83a63a74", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1054369059-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a634a61020b94d62a8fb37a3f96acc10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91733d5a-84", "ovs_interfaceid": "91733d5a-84d6-48df-ba4d-7e362a642282", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 733.641271] env[63531]: DEBUG oslo_vmware.api [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Task: {'id': task-1117605, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.642479] env[63531]: DEBUG nova.network.neutron [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Successfully updated port: c173088d-cfd8-4f05-8936-1d0d5927ae34 {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 733.648940] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:df:a1:31', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '271fe7a0-dfd7-409b-920a-cf83ef1a86a3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '91733d5a-84d6-48df-ba4d-7e362a642282', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 733.657101] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Creating folder: Project (a634a61020b94d62a8fb37a3f96acc10). Parent ref: group-v244585. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 733.658248] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-aecb87e2-9477-49f6-988f-dc7fad102059 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.669520] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Created folder: Project (a634a61020b94d62a8fb37a3f96acc10) in parent group-v244585. [ 733.669520] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Creating folder: Instances. Parent ref: group-v244623. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 733.669520] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-878e8235-d3e3-4abb-aaff-1d2b8f3460a5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.680114] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Created folder: Instances in parent group-v244623. [ 733.680114] env[63531]: DEBUG oslo.service.loopingcall [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 733.680114] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 733.680114] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e47079ff-291a-4b6b-86d9-0a37f8f59020 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.700035] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 733.700035] env[63531]: value = "task-1117608" [ 733.700035] env[63531]: _type = "Task" [ 733.700035] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.704577] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.450s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 733.704577] env[63531]: DEBUG nova.compute.manager [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 733.714027] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.290s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.715308] env[63531]: INFO nova.compute.claims [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 733.734133] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117608, 'name': CreateVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.735696] env[63531]: INFO nova.compute.manager [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: f39c8807-1e15-41cf-899b-a1fbe0695d58] Took 43.43 seconds to build instance. [ 733.743594] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.935169] env[63531]: DEBUG nova.compute.manager [req-c0fc74d4-7ca7-48c6-9de6-470b83bb748c req-87cceb4c-9c54-4bce-99ac-fbb8c0c1e6d5 service nova] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Received event network-changed-91733d5a-84d6-48df-ba4d-7e362a642282 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 733.935261] env[63531]: DEBUG nova.compute.manager [req-c0fc74d4-7ca7-48c6-9de6-470b83bb748c req-87cceb4c-9c54-4bce-99ac-fbb8c0c1e6d5 service nova] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Refreshing instance network info cache due to event network-changed-91733d5a-84d6-48df-ba4d-7e362a642282. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 733.935483] env[63531]: DEBUG oslo_concurrency.lockutils [req-c0fc74d4-7ca7-48c6-9de6-470b83bb748c req-87cceb4c-9c54-4bce-99ac-fbb8c0c1e6d5 service nova] Acquiring lock "refresh_cache-c3920e60-b27f-496b-b616-afb23ab0bc26" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 733.935630] env[63531]: DEBUG oslo_concurrency.lockutils [req-c0fc74d4-7ca7-48c6-9de6-470b83bb748c req-87cceb4c-9c54-4bce-99ac-fbb8c0c1e6d5 service nova] Acquired lock "refresh_cache-c3920e60-b27f-496b-b616-afb23ab0bc26" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 733.935794] env[63531]: DEBUG nova.network.neutron [req-c0fc74d4-7ca7-48c6-9de6-470b83bb748c req-87cceb4c-9c54-4bce-99ac-fbb8c0c1e6d5 service nova] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Refreshing network info cache for port 91733d5a-84d6-48df-ba4d-7e362a642282 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 734.139560] env[63531]: DEBUG oslo_vmware.api [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Task: {'id': task-1117605, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.158559] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Acquiring lock "refresh_cache-410ddf7a-e45b-4769-a3db-5363ce2096a7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 734.158832] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Acquired lock "refresh_cache-410ddf7a-e45b-4769-a3db-5363ce2096a7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 734.159067] env[63531]: DEBUG nova.network.neutron [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 734.214892] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117608, 'name': CreateVM_Task, 'duration_secs': 0.415105} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.215295] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 734.215958] env[63531]: DEBUG oslo_concurrency.lockutils [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 734.216141] env[63531]: DEBUG oslo_concurrency.lockutils [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 734.216557] env[63531]: DEBUG oslo_concurrency.lockutils [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 734.216727] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c53764a0-2142-4cbc-8b63-602b886223a0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.225331] env[63531]: DEBUG nova.compute.utils [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 734.226741] env[63531]: DEBUG oslo_vmware.api [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Waiting for the task: (returnval){ [ 734.226741] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52e0c14e-c65e-5912-a475-0aef5e5448e5" [ 734.226741] env[63531]: _type = "Task" [ 734.226741] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.228054] env[63531]: DEBUG nova.compute.manager [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 734.228249] env[63531]: DEBUG nova.network.neutron [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 734.240244] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4422eb98-0c9a-4f30-af98-d51dabfb2807 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Lock "f39c8807-1e15-41cf-899b-a1fbe0695d58" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 128.070s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 734.241114] env[63531]: DEBUG oslo_vmware.api [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52e0c14e-c65e-5912-a475-0aef5e5448e5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.277408] env[63531]: DEBUG nova.policy [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7965dce40de84f8083d57a8d51d4a063', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f727ddb992dc4c06954f25f6a1613b4a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 734.544172] env[63531]: DEBUG nova.network.neutron [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Successfully created port: a70530fb-7805-40a0-bea5-a2a666dc679d {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 734.638011] env[63531]: DEBUG oslo_vmware.api [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Task: {'id': task-1117605, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.602392} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.640794] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 5e9042b3-4dc2-4fa3-a664-c4b49a22e400/5e9042b3-4dc2-4fa3-a664-c4b49a22e400.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 734.641042] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 734.641913] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b6851a7a-0404-4bdb-a5ca-2f3a393e9945 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.650938] env[63531]: DEBUG oslo_vmware.api [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Waiting for the task: (returnval){ [ 734.650938] env[63531]: value = "task-1117609" [ 734.650938] env[63531]: _type = "Task" [ 734.650938] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.660259] env[63531]: DEBUG oslo_vmware.api [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Task: {'id': task-1117609, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.717226] env[63531]: DEBUG oslo_concurrency.lockutils [None req-db4fd4eb-d9b5-45f3-8b8c-2517c717ed2c tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Acquiring lock "c3d597b4-1dc3-48d8-9bee-e73c8929181b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 734.717226] env[63531]: DEBUG oslo_concurrency.lockutils [None req-db4fd4eb-d9b5-45f3-8b8c-2517c717ed2c tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Lock "c3d597b4-1dc3-48d8-9bee-e73c8929181b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 734.717226] env[63531]: DEBUG oslo_concurrency.lockutils [None req-db4fd4eb-d9b5-45f3-8b8c-2517c717ed2c tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Acquiring lock "c3d597b4-1dc3-48d8-9bee-e73c8929181b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 734.717533] env[63531]: DEBUG oslo_concurrency.lockutils [None req-db4fd4eb-d9b5-45f3-8b8c-2517c717ed2c tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Lock "c3d597b4-1dc3-48d8-9bee-e73c8929181b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 734.717758] env[63531]: DEBUG oslo_concurrency.lockutils [None req-db4fd4eb-d9b5-45f3-8b8c-2517c717ed2c tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Lock "c3d597b4-1dc3-48d8-9bee-e73c8929181b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 734.720090] env[63531]: INFO nova.compute.manager [None req-db4fd4eb-d9b5-45f3-8b8c-2517c717ed2c tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Terminating instance [ 734.722009] env[63531]: DEBUG nova.compute.manager [None req-db4fd4eb-d9b5-45f3-8b8c-2517c717ed2c tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 734.722188] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-db4fd4eb-d9b5-45f3-8b8c-2517c717ed2c tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 734.723019] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22cd5c67-7006-4e78-a709-a2de6590ea2c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.727977] env[63531]: DEBUG nova.compute.manager [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 734.739055] env[63531]: DEBUG nova.network.neutron [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 734.741209] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-db4fd4eb-d9b5-45f3-8b8c-2517c717ed2c tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 734.745070] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-238be2a6-b295-43ea-a219-8529180017ad {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.747850] env[63531]: DEBUG nova.compute.manager [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] [instance: dccb2335-4220-4570-861b-16ff2d8f5a85] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 734.759154] env[63531]: DEBUG oslo_vmware.api [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52e0c14e-c65e-5912-a475-0aef5e5448e5, 'name': SearchDatastore_Task, 'duration_secs': 0.01816} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.761308] env[63531]: DEBUG oslo_concurrency.lockutils [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 734.761731] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 734.761843] env[63531]: DEBUG oslo_concurrency.lockutils [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 734.761936] env[63531]: DEBUG oslo_concurrency.lockutils [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 734.762338] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 734.762732] env[63531]: DEBUG oslo_vmware.api [None req-db4fd4eb-d9b5-45f3-8b8c-2517c717ed2c tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Waiting for the task: (returnval){ [ 734.762732] env[63531]: value = "task-1117610" [ 734.762732] env[63531]: _type = "Task" [ 734.762732] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.767406] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-98cf7ab9-f703-4b28-9638-98bd3513f05b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.781757] env[63531]: DEBUG oslo_vmware.api [None req-db4fd4eb-d9b5-45f3-8b8c-2517c717ed2c tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': task-1117610, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.783260] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 734.783445] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 734.784179] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2c4e74a9-fd51-46bd-a3bc-979a41adb21c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.794669] env[63531]: DEBUG oslo_vmware.api [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Waiting for the task: (returnval){ [ 734.794669] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]529ecf4c-4f28-4caa-13ec-de9d69d864ec" [ 734.794669] env[63531]: _type = "Task" [ 734.794669] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.805324] env[63531]: DEBUG oslo_vmware.api [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]529ecf4c-4f28-4caa-13ec-de9d69d864ec, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.891099] env[63531]: DEBUG nova.network.neutron [req-c0fc74d4-7ca7-48c6-9de6-470b83bb748c req-87cceb4c-9c54-4bce-99ac-fbb8c0c1e6d5 service nova] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Updated VIF entry in instance network info cache for port 91733d5a-84d6-48df-ba4d-7e362a642282. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 734.892016] env[63531]: DEBUG nova.network.neutron [req-c0fc74d4-7ca7-48c6-9de6-470b83bb748c req-87cceb4c-9c54-4bce-99ac-fbb8c0c1e6d5 service nova] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Updating instance_info_cache with network_info: [{"id": "91733d5a-84d6-48df-ba4d-7e362a642282", "address": "fa:16:3e:df:a1:31", "network": {"id": "d28b2087-329c-4917-aa6e-83ab83a63a74", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1054369059-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a634a61020b94d62a8fb37a3f96acc10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91733d5a-84", "ovs_interfaceid": "91733d5a-84d6-48df-ba4d-7e362a642282", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.006937] env[63531]: DEBUG nova.compute.manager [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 735.007823] env[63531]: DEBUG nova.virt.hardware [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 735.008189] env[63531]: DEBUG nova.virt.hardware [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 735.008464] env[63531]: DEBUG nova.virt.hardware [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 735.008785] env[63531]: DEBUG nova.virt.hardware [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 735.009066] env[63531]: DEBUG nova.virt.hardware [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 735.009329] env[63531]: DEBUG nova.virt.hardware [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 735.009667] env[63531]: DEBUG nova.virt.hardware [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 735.009939] env[63531]: DEBUG nova.virt.hardware [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 735.010251] env[63531]: DEBUG nova.virt.hardware [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 735.010553] env[63531]: DEBUG nova.virt.hardware [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 735.010881] env[63531]: DEBUG nova.virt.hardware [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 735.017765] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddcde81c-1710-4802-b761-8b09fb6cc9b9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.029122] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9d5f40a-2421-4621-bf0c-336b3c1162f9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.050082] env[63531]: DEBUG nova.network.neutron [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Updating instance_info_cache with network_info: [{"id": "c173088d-cfd8-4f05-8936-1d0d5927ae34", "address": "fa:16:3e:24:0f:e5", "network": {"id": "88ee583d-32f8-4f4e-b02c-392260bab32b", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-2110978471-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fed6b4862d76469b87ecf16dc6dce177", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc173088d-cf", "ovs_interfaceid": "c173088d-cfd8-4f05-8936-1d0d5927ae34", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.146874] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74619968-d959-4df0-84c2-0ee1c112c38b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.157071] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71f26fbf-08b5-4021-8b7f-9ec586ec065e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.164877] env[63531]: DEBUG oslo_vmware.api [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Task: {'id': task-1117609, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.188404} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.189904] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 735.190882] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4b98124-8d3b-43f4-b031-ba4da29d45e3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.193759] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-551b0fbb-8fec-4d50-bbdf-006100b1f613 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.209716] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f9d10eb-65b6-4b80-b8c4-62e427471358 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.221652] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Reconfiguring VM instance instance-00000027 to attach disk [datastore1] 5e9042b3-4dc2-4fa3-a664-c4b49a22e400/5e9042b3-4dc2-4fa3-a664-c4b49a22e400.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 735.221901] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-54efe6f0-8293-4fdb-867c-b4f6c4feac2f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.247519] env[63531]: DEBUG nova.compute.provider_tree [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 735.249634] env[63531]: DEBUG oslo_vmware.api [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Waiting for the task: (returnval){ [ 735.249634] env[63531]: value = "task-1117611" [ 735.249634] env[63531]: _type = "Task" [ 735.249634] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.260892] env[63531]: DEBUG oslo_vmware.api [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Task: {'id': task-1117611, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.270746] env[63531]: DEBUG oslo_concurrency.lockutils [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 735.280106] env[63531]: DEBUG oslo_vmware.api [None req-db4fd4eb-d9b5-45f3-8b8c-2517c717ed2c tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': task-1117610, 'name': PowerOffVM_Task, 'duration_secs': 0.210951} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.280435] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-db4fd4eb-d9b5-45f3-8b8c-2517c717ed2c tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 735.280568] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-db4fd4eb-d9b5-45f3-8b8c-2517c717ed2c tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 735.280792] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dfc9c368-5bac-4b94-be39-2d51a9521c86 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.304484] env[63531]: DEBUG oslo_vmware.api [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]529ecf4c-4f28-4caa-13ec-de9d69d864ec, 'name': SearchDatastore_Task, 'duration_secs': 0.015051} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.304991] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5c4f687f-3921-44bd-9d0e-9ddf87dcf920 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.310695] env[63531]: DEBUG oslo_vmware.api [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Waiting for the task: (returnval){ [ 735.310695] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]528bec1f-1689-5f6e-ac31-6fea2994fe22" [ 735.310695] env[63531]: _type = "Task" [ 735.310695] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.318560] env[63531]: DEBUG oslo_vmware.api [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]528bec1f-1689-5f6e-ac31-6fea2994fe22, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.350940] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-db4fd4eb-d9b5-45f3-8b8c-2517c717ed2c tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 735.351135] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-db4fd4eb-d9b5-45f3-8b8c-2517c717ed2c tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Deleting contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 735.351371] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-db4fd4eb-d9b5-45f3-8b8c-2517c717ed2c tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Deleting the datastore file [datastore2] c3d597b4-1dc3-48d8-9bee-e73c8929181b {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 735.351632] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c8211897-fb59-4760-a092-69ebc611f0e1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.357812] env[63531]: DEBUG oslo_vmware.api [None req-db4fd4eb-d9b5-45f3-8b8c-2517c717ed2c tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Waiting for the task: (returnval){ [ 735.357812] env[63531]: value = "task-1117613" [ 735.357812] env[63531]: _type = "Task" [ 735.357812] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.365557] env[63531]: DEBUG oslo_vmware.api [None req-db4fd4eb-d9b5-45f3-8b8c-2517c717ed2c tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': task-1117613, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.396627] env[63531]: DEBUG oslo_concurrency.lockutils [req-c0fc74d4-7ca7-48c6-9de6-470b83bb748c req-87cceb4c-9c54-4bce-99ac-fbb8c0c1e6d5 service nova] Releasing lock "refresh_cache-c3920e60-b27f-496b-b616-afb23ab0bc26" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 735.554923] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Releasing lock "refresh_cache-410ddf7a-e45b-4769-a3db-5363ce2096a7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 735.555354] env[63531]: DEBUG nova.compute.manager [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Instance network_info: |[{"id": "c173088d-cfd8-4f05-8936-1d0d5927ae34", "address": "fa:16:3e:24:0f:e5", "network": {"id": "88ee583d-32f8-4f4e-b02c-392260bab32b", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-2110978471-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fed6b4862d76469b87ecf16dc6dce177", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc173088d-cf", "ovs_interfaceid": "c173088d-cfd8-4f05-8936-1d0d5927ae34", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 735.555879] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:24:0f:e5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4a8f5363-be3a-4f92-9ccf-33bb0c8113b3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c173088d-cfd8-4f05-8936-1d0d5927ae34', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 735.564276] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Creating folder: Project (fed6b4862d76469b87ecf16dc6dce177). Parent ref: group-v244585. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 735.564601] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8d0d8c35-f140-4537-9793-684900268635 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.578878] env[63531]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 735.579116] env[63531]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=63531) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 735.579511] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Folder already exists: Project (fed6b4862d76469b87ecf16dc6dce177). Parent ref: group-v244585. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 735.579770] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Creating folder: Instances. Parent ref: group-v244598. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 735.580073] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-178f2da4-b8ee-45b5-9a6d-16f8b7a01766 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.594012] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Created folder: Instances in parent group-v244598. [ 735.594012] env[63531]: DEBUG oslo.service.loopingcall [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 735.594012] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 735.594012] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5d478a58-dcfd-4f6c-b68b-78f68eae33c4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.618108] env[63531]: DEBUG nova.compute.manager [req-a2d1b6f4-c518-4476-ae17-e107b7a1014f req-958d0929-0a1a-4af9-8096-177e28ea0ce0 service nova] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Received event network-changed-c173088d-cfd8-4f05-8936-1d0d5927ae34 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 735.618362] env[63531]: DEBUG nova.compute.manager [req-a2d1b6f4-c518-4476-ae17-e107b7a1014f req-958d0929-0a1a-4af9-8096-177e28ea0ce0 service nova] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Refreshing instance network info cache due to event network-changed-c173088d-cfd8-4f05-8936-1d0d5927ae34. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 735.618588] env[63531]: DEBUG oslo_concurrency.lockutils [req-a2d1b6f4-c518-4476-ae17-e107b7a1014f req-958d0929-0a1a-4af9-8096-177e28ea0ce0 service nova] Acquiring lock "refresh_cache-410ddf7a-e45b-4769-a3db-5363ce2096a7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 735.618732] env[63531]: DEBUG oslo_concurrency.lockutils [req-a2d1b6f4-c518-4476-ae17-e107b7a1014f req-958d0929-0a1a-4af9-8096-177e28ea0ce0 service nova] Acquired lock "refresh_cache-410ddf7a-e45b-4769-a3db-5363ce2096a7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 735.618902] env[63531]: DEBUG nova.network.neutron [req-a2d1b6f4-c518-4476-ae17-e107b7a1014f req-958d0929-0a1a-4af9-8096-177e28ea0ce0 service nova] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Refreshing network info cache for port c173088d-cfd8-4f05-8936-1d0d5927ae34 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 735.623735] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 735.623735] env[63531]: value = "task-1117616" [ 735.623735] env[63531]: _type = "Task" [ 735.623735] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.631518] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117616, 'name': CreateVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.750307] env[63531]: DEBUG nova.compute.manager [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 735.753666] env[63531]: DEBUG nova.scheduler.client.report [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 735.766526] env[63531]: DEBUG oslo_vmware.api [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Task: {'id': task-1117611, 'name': ReconfigVM_Task, 'duration_secs': 0.310273} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.766821] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Reconfigured VM instance instance-00000027 to attach disk [datastore1] 5e9042b3-4dc2-4fa3-a664-c4b49a22e400/5e9042b3-4dc2-4fa3-a664-c4b49a22e400.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 735.767482] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-44543c83-4310-495c-8543-de6817c0f0eb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.776819] env[63531]: DEBUG oslo_vmware.api [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Waiting for the task: (returnval){ [ 735.776819] env[63531]: value = "task-1117617" [ 735.776819] env[63531]: _type = "Task" [ 735.776819] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.779505] env[63531]: DEBUG nova.virt.hardware [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 735.779883] env[63531]: DEBUG nova.virt.hardware [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 735.780163] env[63531]: DEBUG nova.virt.hardware [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 735.780494] env[63531]: DEBUG nova.virt.hardware [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 735.780719] env[63531]: DEBUG nova.virt.hardware [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 735.780913] env[63531]: DEBUG nova.virt.hardware [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 735.781164] env[63531]: DEBUG nova.virt.hardware [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 735.781362] env[63531]: DEBUG nova.virt.hardware [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 735.781534] env[63531]: DEBUG nova.virt.hardware [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 735.781828] env[63531]: DEBUG nova.virt.hardware [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 735.781945] env[63531]: DEBUG nova.virt.hardware [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 735.783206] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5723d521-ecb6-4931-80fb-aea84db3b30f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.794431] env[63531]: DEBUG oslo_vmware.api [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Task: {'id': task-1117617, 'name': Rename_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.797468] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02dafe31-9996-4596-82e3-b2d2304adb53 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.820536] env[63531]: DEBUG oslo_vmware.api [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]528bec1f-1689-5f6e-ac31-6fea2994fe22, 'name': SearchDatastore_Task, 'duration_secs': 0.010638} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.820708] env[63531]: DEBUG oslo_concurrency.lockutils [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 735.821361] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] c3920e60-b27f-496b-b616-afb23ab0bc26/c3920e60-b27f-496b-b616-afb23ab0bc26.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 735.821361] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6df3e7a1-cd95-4471-ae50-4c2544b5ead1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.827864] env[63531]: DEBUG oslo_vmware.api [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Waiting for the task: (returnval){ [ 735.827864] env[63531]: value = "task-1117618" [ 735.827864] env[63531]: _type = "Task" [ 735.827864] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.837879] env[63531]: DEBUG oslo_vmware.api [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Task: {'id': task-1117618, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.868955] env[63531]: DEBUG oslo_vmware.api [None req-db4fd4eb-d9b5-45f3-8b8c-2517c717ed2c tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': task-1117613, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.169273} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.869276] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-db4fd4eb-d9b5-45f3-8b8c-2517c717ed2c tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 735.869529] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-db4fd4eb-d9b5-45f3-8b8c-2517c717ed2c tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Deleted contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 735.869652] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-db4fd4eb-d9b5-45f3-8b8c-2517c717ed2c tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 735.869835] env[63531]: INFO nova.compute.manager [None req-db4fd4eb-d9b5-45f3-8b8c-2517c717ed2c tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Took 1.15 seconds to destroy the instance on the hypervisor. [ 735.870121] env[63531]: DEBUG oslo.service.loopingcall [None req-db4fd4eb-d9b5-45f3-8b8c-2517c717ed2c tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 735.870327] env[63531]: DEBUG nova.compute.manager [-] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 735.870428] env[63531]: DEBUG nova.network.neutron [-] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 736.085230] env[63531]: DEBUG nova.network.neutron [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Successfully updated port: a70530fb-7805-40a0-bea5-a2a666dc679d {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 736.135220] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117616, 'name': CreateVM_Task, 'duration_secs': 0.375561} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.135220] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 736.136301] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'mount_device': '/dev/sda', 'attachment_id': '95f063b2-3b3b-458f-a9f8-d98f48414b1a', 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244601', 'volume_id': '2a24c282-f681-40bf-a19c-5c711d342687', 'name': 'volume-2a24c282-f681-40bf-a19c-5c711d342687', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '410ddf7a-e45b-4769-a3db-5363ce2096a7', 'attached_at': '', 'detached_at': '', 'volume_id': '2a24c282-f681-40bf-a19c-5c711d342687', 'serial': '2a24c282-f681-40bf-a19c-5c711d342687'}, 'disk_bus': None, 'guest_format': None, 'delete_on_termination': True, 'boot_index': 0, 'device_type': None, 'volume_type': None}], 'swap': None} {{(pid=63531) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 736.136301] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Root volume attach. Driver type: vmdk {{(pid=63531) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 736.137021] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-055690ff-6888-42ae-bdb8-b7e31a3f4371 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.147344] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a63e054-64e4-4444-adb4-f433c556963c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.155105] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a3bdd4c-7914-4493-b713-008b9947a811 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.165736] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-8f7da269-4932-4d2e-8e1b-a87d58cb7bf2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.175361] env[63531]: DEBUG oslo_vmware.api [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Waiting for the task: (returnval){ [ 736.175361] env[63531]: value = "task-1117619" [ 736.175361] env[63531]: _type = "Task" [ 736.175361] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.188763] env[63531]: DEBUG oslo_vmware.api [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Task: {'id': task-1117619, 'name': RelocateVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.262244] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.550s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.262775] env[63531]: DEBUG nova.compute.manager [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 736.266009] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.674s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 736.267673] env[63531]: INFO nova.compute.claims [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] [instance: 0af99240-fa7a-4eed-a729-a5ae98c41cf9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 736.295491] env[63531]: DEBUG oslo_vmware.api [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Task: {'id': task-1117617, 'name': Rename_Task, 'duration_secs': 0.204369} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.295809] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 736.296223] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-12e11052-980e-4d97-9f6b-3227df94972c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.311635] env[63531]: DEBUG oslo_vmware.api [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Waiting for the task: (returnval){ [ 736.311635] env[63531]: value = "task-1117620" [ 736.311635] env[63531]: _type = "Task" [ 736.311635] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.322245] env[63531]: DEBUG oslo_vmware.api [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Task: {'id': task-1117620, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.328075] env[63531]: DEBUG nova.compute.manager [req-7d476ad1-45c2-4758-aa00-a73faa9cfcac req-eb3bb877-c6a2-4967-941f-c5fb06dfa315 service nova] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Received event network-vif-deleted-0c5ad75f-1b78-47fd-bd2e-331a7ae0732a {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 736.328321] env[63531]: INFO nova.compute.manager [req-7d476ad1-45c2-4758-aa00-a73faa9cfcac req-eb3bb877-c6a2-4967-941f-c5fb06dfa315 service nova] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Neutron deleted interface 0c5ad75f-1b78-47fd-bd2e-331a7ae0732a; detaching it from the instance and deleting it from the info cache [ 736.328812] env[63531]: DEBUG nova.network.neutron [req-7d476ad1-45c2-4758-aa00-a73faa9cfcac req-eb3bb877-c6a2-4967-941f-c5fb06dfa315 service nova] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 736.340019] env[63531]: DEBUG oslo_vmware.api [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Task: {'id': task-1117618, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.387231] env[63531]: DEBUG nova.network.neutron [req-a2d1b6f4-c518-4476-ae17-e107b7a1014f req-958d0929-0a1a-4af9-8096-177e28ea0ce0 service nova] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Updated VIF entry in instance network info cache for port c173088d-cfd8-4f05-8936-1d0d5927ae34. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 736.387594] env[63531]: DEBUG nova.network.neutron [req-a2d1b6f4-c518-4476-ae17-e107b7a1014f req-958d0929-0a1a-4af9-8096-177e28ea0ce0 service nova] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Updating instance_info_cache with network_info: [{"id": "c173088d-cfd8-4f05-8936-1d0d5927ae34", "address": "fa:16:3e:24:0f:e5", "network": {"id": "88ee583d-32f8-4f4e-b02c-392260bab32b", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-2110978471-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fed6b4862d76469b87ecf16dc6dce177", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc173088d-cf", "ovs_interfaceid": "c173088d-cfd8-4f05-8936-1d0d5927ae34", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 736.586779] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Acquiring lock "refresh_cache-1ca7b851-2a06-4181-8271-58aafcd322d6" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 736.587010] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Acquired lock "refresh_cache-1ca7b851-2a06-4181-8271-58aafcd322d6" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 736.587188] env[63531]: DEBUG nova.network.neutron [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 736.685728] env[63531]: DEBUG oslo_vmware.api [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Task: {'id': task-1117619, 'name': RelocateVM_Task, 'duration_secs': 0.029544} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.686045] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Volume attach. Driver type: vmdk {{(pid=63531) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 736.686258] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244601', 'volume_id': '2a24c282-f681-40bf-a19c-5c711d342687', 'name': 'volume-2a24c282-f681-40bf-a19c-5c711d342687', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '410ddf7a-e45b-4769-a3db-5363ce2096a7', 'attached_at': '', 'detached_at': '', 'volume_id': '2a24c282-f681-40bf-a19c-5c711d342687', 'serial': '2a24c282-f681-40bf-a19c-5c711d342687'} {{(pid=63531) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 736.687110] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90aca379-b060-4e2e-9fc7-7f064e23b457 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.703027] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99d031a4-e44b-44ab-8c3a-c4b80a3701a5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.724145] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Reconfiguring VM instance instance-00000029 to attach disk [datastore1] volume-2a24c282-f681-40bf-a19c-5c711d342687/volume-2a24c282-f681-40bf-a19c-5c711d342687.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 736.724771] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1af97646-e0c4-4291-832d-3a39b0cce64f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.743513] env[63531]: DEBUG oslo_vmware.api [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Waiting for the task: (returnval){ [ 736.743513] env[63531]: value = "task-1117621" [ 736.743513] env[63531]: _type = "Task" [ 736.743513] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.751941] env[63531]: DEBUG oslo_vmware.api [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Task: {'id': task-1117621, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.757529] env[63531]: DEBUG nova.network.neutron [-] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 736.771731] env[63531]: DEBUG nova.compute.utils [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 736.774758] env[63531]: DEBUG nova.compute.manager [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 736.774933] env[63531]: DEBUG nova.network.neutron [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 736.819626] env[63531]: DEBUG nova.policy [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '69131938118a405b9bb362be787a3f14', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '824ed405cc6e4868a70e82489e05e8ac', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 736.825260] env[63531]: DEBUG oslo_vmware.api [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Task: {'id': task-1117620, 'name': PowerOnVM_Task} progress is 92%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.834126] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5dc725c4-509c-445a-b151-e5001fe90ded {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.839144] env[63531]: DEBUG oslo_vmware.api [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Task: {'id': task-1117618, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.588415} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.839722] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] c3920e60-b27f-496b-b616-afb23ab0bc26/c3920e60-b27f-496b-b616-afb23ab0bc26.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 736.839944] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 736.840231] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-55080d20-a465-4002-a41b-3764a7451667 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.845622] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e833d6cb-3db0-4639-a22e-e4571089b93d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.857369] env[63531]: DEBUG oslo_vmware.api [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Waiting for the task: (returnval){ [ 736.857369] env[63531]: value = "task-1117622" [ 736.857369] env[63531]: _type = "Task" [ 736.857369] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.865966] env[63531]: DEBUG oslo_vmware.api [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Task: {'id': task-1117622, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.872108] env[63531]: DEBUG nova.compute.manager [req-7d476ad1-45c2-4758-aa00-a73faa9cfcac req-eb3bb877-c6a2-4967-941f-c5fb06dfa315 service nova] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Detach interface failed, port_id=0c5ad75f-1b78-47fd-bd2e-331a7ae0732a, reason: Instance c3d597b4-1dc3-48d8-9bee-e73c8929181b could not be found. {{(pid=63531) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 736.889933] env[63531]: DEBUG oslo_concurrency.lockutils [req-a2d1b6f4-c518-4476-ae17-e107b7a1014f req-958d0929-0a1a-4af9-8096-177e28ea0ce0 service nova] Releasing lock "refresh_cache-410ddf7a-e45b-4769-a3db-5363ce2096a7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 737.069474] env[63531]: DEBUG nova.network.neutron [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Successfully created port: 6102475b-8e43-4b77-85c3-5b574383c06d {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 737.136825] env[63531]: DEBUG nova.network.neutron [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 737.253387] env[63531]: DEBUG oslo_vmware.api [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Task: {'id': task-1117621, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.261095] env[63531]: INFO nova.compute.manager [-] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Took 1.39 seconds to deallocate network for instance. [ 737.280552] env[63531]: DEBUG nova.compute.manager [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 737.326660] env[63531]: DEBUG oslo_vmware.api [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Task: {'id': task-1117620, 'name': PowerOnVM_Task, 'duration_secs': 0.555189} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.328640] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 737.328854] env[63531]: INFO nova.compute.manager [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Took 9.03 seconds to spawn the instance on the hypervisor. [ 737.329041] env[63531]: DEBUG nova.compute.manager [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 737.332098] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eadd36d1-8256-416f-8cd4-2a54d05fe433 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.370516] env[63531]: DEBUG oslo_vmware.api [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Task: {'id': task-1117622, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.310362} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.370713] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 737.371630] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50e391ac-6a7a-49bd-96d6-ac8ffd6f409a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.393405] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Reconfiguring VM instance instance-00000028 to attach disk [datastore1] c3920e60-b27f-496b-b616-afb23ab0bc26/c3920e60-b27f-496b-b616-afb23ab0bc26.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 737.394430] env[63531]: DEBUG nova.network.neutron [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Updating instance_info_cache with network_info: [{"id": "a70530fb-7805-40a0-bea5-a2a666dc679d", "address": "fa:16:3e:78:c3:90", "network": {"id": "73e6d3ed-2016-498a-bda6-fd7dfbff93da", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1501868918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f727ddb992dc4c06954f25f6a1613b4a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c20f5114-0866-45b3-9a7c-62f113ff83fa", "external-id": "nsx-vlan-transportzone-47", "segmentation_id": 47, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa70530fb-78", "ovs_interfaceid": "a70530fb-7805-40a0-bea5-a2a666dc679d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 737.398694] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0df42d0c-554d-496e-bb34-7edf75dc241e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.419931] env[63531]: DEBUG oslo_vmware.api [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Waiting for the task: (returnval){ [ 737.419931] env[63531]: value = "task-1117623" [ 737.419931] env[63531]: _type = "Task" [ 737.419931] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.431031] env[63531]: DEBUG oslo_vmware.api [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Task: {'id': task-1117623, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.676136] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5720b84a-97e4-409b-8daf-e85d891ffa7c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.684210] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca93a84f-004b-43ea-94bc-75c1c243bed9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.690023] env[63531]: DEBUG nova.compute.manager [req-77854249-de53-4228-aaf9-fe40b7eabc74 req-4476da96-3d24-45c3-a174-e5f7fc0f98d4 service nova] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Received event network-vif-plugged-a70530fb-7805-40a0-bea5-a2a666dc679d {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 737.690023] env[63531]: DEBUG oslo_concurrency.lockutils [req-77854249-de53-4228-aaf9-fe40b7eabc74 req-4476da96-3d24-45c3-a174-e5f7fc0f98d4 service nova] Acquiring lock "1ca7b851-2a06-4181-8271-58aafcd322d6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 737.690295] env[63531]: DEBUG oslo_concurrency.lockutils [req-77854249-de53-4228-aaf9-fe40b7eabc74 req-4476da96-3d24-45c3-a174-e5f7fc0f98d4 service nova] Lock "1ca7b851-2a06-4181-8271-58aafcd322d6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 737.690510] env[63531]: DEBUG oslo_concurrency.lockutils [req-77854249-de53-4228-aaf9-fe40b7eabc74 req-4476da96-3d24-45c3-a174-e5f7fc0f98d4 service nova] Lock "1ca7b851-2a06-4181-8271-58aafcd322d6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 737.690718] env[63531]: DEBUG nova.compute.manager [req-77854249-de53-4228-aaf9-fe40b7eabc74 req-4476da96-3d24-45c3-a174-e5f7fc0f98d4 service nova] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] No waiting events found dispatching network-vif-plugged-a70530fb-7805-40a0-bea5-a2a666dc679d {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 737.690916] env[63531]: WARNING nova.compute.manager [req-77854249-de53-4228-aaf9-fe40b7eabc74 req-4476da96-3d24-45c3-a174-e5f7fc0f98d4 service nova] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Received unexpected event network-vif-plugged-a70530fb-7805-40a0-bea5-a2a666dc679d for instance with vm_state building and task_state spawning. [ 737.691112] env[63531]: DEBUG nova.compute.manager [req-77854249-de53-4228-aaf9-fe40b7eabc74 req-4476da96-3d24-45c3-a174-e5f7fc0f98d4 service nova] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Received event network-changed-a70530fb-7805-40a0-bea5-a2a666dc679d {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 737.691299] env[63531]: DEBUG nova.compute.manager [req-77854249-de53-4228-aaf9-fe40b7eabc74 req-4476da96-3d24-45c3-a174-e5f7fc0f98d4 service nova] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Refreshing instance network info cache due to event network-changed-a70530fb-7805-40a0-bea5-a2a666dc679d. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 737.691492] env[63531]: DEBUG oslo_concurrency.lockutils [req-77854249-de53-4228-aaf9-fe40b7eabc74 req-4476da96-3d24-45c3-a174-e5f7fc0f98d4 service nova] Acquiring lock "refresh_cache-1ca7b851-2a06-4181-8271-58aafcd322d6" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 737.719062] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7baa6d3-4e4c-4288-b2a5-becc0fe4d286 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.726603] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10e4ff41-e314-44c8-bff1-9cb3ad0a0a33 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.745386] env[63531]: DEBUG nova.compute.provider_tree [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 737.756335] env[63531]: DEBUG oslo_vmware.api [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Task: {'id': task-1117621, 'name': ReconfigVM_Task, 'duration_secs': 0.934447} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.756643] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Reconfigured VM instance instance-00000029 to attach disk [datastore1] volume-2a24c282-f681-40bf-a19c-5c711d342687/volume-2a24c282-f681-40bf-a19c-5c711d342687.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 737.765128] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fb96f5de-d2a9-46c2-9f88-03496c0668a3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.781655] env[63531]: DEBUG oslo_concurrency.lockutils [None req-db4fd4eb-d9b5-45f3-8b8c-2517c717ed2c tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 737.794609] env[63531]: DEBUG oslo_vmware.api [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Waiting for the task: (returnval){ [ 737.794609] env[63531]: value = "task-1117624" [ 737.794609] env[63531]: _type = "Task" [ 737.794609] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.804407] env[63531]: DEBUG oslo_vmware.api [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Task: {'id': task-1117624, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.850895] env[63531]: INFO nova.compute.manager [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Took 30.25 seconds to build instance. [ 737.915221] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Releasing lock "refresh_cache-1ca7b851-2a06-4181-8271-58aafcd322d6" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 737.918020] env[63531]: DEBUG nova.compute.manager [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Instance network_info: |[{"id": "a70530fb-7805-40a0-bea5-a2a666dc679d", "address": "fa:16:3e:78:c3:90", "network": {"id": "73e6d3ed-2016-498a-bda6-fd7dfbff93da", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1501868918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f727ddb992dc4c06954f25f6a1613b4a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c20f5114-0866-45b3-9a7c-62f113ff83fa", "external-id": "nsx-vlan-transportzone-47", "segmentation_id": 47, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa70530fb-78", "ovs_interfaceid": "a70530fb-7805-40a0-bea5-a2a666dc679d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 737.918020] env[63531]: DEBUG oslo_concurrency.lockutils [req-77854249-de53-4228-aaf9-fe40b7eabc74 req-4476da96-3d24-45c3-a174-e5f7fc0f98d4 service nova] Acquired lock "refresh_cache-1ca7b851-2a06-4181-8271-58aafcd322d6" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 737.918321] env[63531]: DEBUG nova.network.neutron [req-77854249-de53-4228-aaf9-fe40b7eabc74 req-4476da96-3d24-45c3-a174-e5f7fc0f98d4 service nova] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Refreshing network info cache for port a70530fb-7805-40a0-bea5-a2a666dc679d {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 737.918321] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:78:c3:90', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c20f5114-0866-45b3-9a7c-62f113ff83fa', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a70530fb-7805-40a0-bea5-a2a666dc679d', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 737.926782] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Creating folder: Project (f727ddb992dc4c06954f25f6a1613b4a). Parent ref: group-v244585. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 737.930057] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b1a4b3af-906b-490c-8c5b-677488c7609a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.945562] env[63531]: DEBUG oslo_vmware.api [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Task: {'id': task-1117623, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.946819] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Created folder: Project (f727ddb992dc4c06954f25f6a1613b4a) in parent group-v244585. [ 737.947015] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Creating folder: Instances. Parent ref: group-v244628. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 737.947254] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8a9a0380-5b7d-411f-b013-07ad589087e8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.955497] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Created folder: Instances in parent group-v244628. [ 737.955718] env[63531]: DEBUG oslo.service.loopingcall [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 737.955921] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 737.956097] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d86d1c7e-07ce-4913-bde2-018ed377cb78 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.976740] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 737.976740] env[63531]: value = "task-1117627" [ 737.976740] env[63531]: _type = "Task" [ 737.976740] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.989125] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117627, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.251759] env[63531]: DEBUG nova.scheduler.client.report [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 738.293200] env[63531]: DEBUG nova.network.neutron [req-77854249-de53-4228-aaf9-fe40b7eabc74 req-4476da96-3d24-45c3-a174-e5f7fc0f98d4 service nova] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Updated VIF entry in instance network info cache for port a70530fb-7805-40a0-bea5-a2a666dc679d. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 738.293549] env[63531]: DEBUG nova.network.neutron [req-77854249-de53-4228-aaf9-fe40b7eabc74 req-4476da96-3d24-45c3-a174-e5f7fc0f98d4 service nova] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Updating instance_info_cache with network_info: [{"id": "a70530fb-7805-40a0-bea5-a2a666dc679d", "address": "fa:16:3e:78:c3:90", "network": {"id": "73e6d3ed-2016-498a-bda6-fd7dfbff93da", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1501868918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f727ddb992dc4c06954f25f6a1613b4a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c20f5114-0866-45b3-9a7c-62f113ff83fa", "external-id": "nsx-vlan-transportzone-47", "segmentation_id": 47, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa70530fb-78", "ovs_interfaceid": "a70530fb-7805-40a0-bea5-a2a666dc679d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 738.296262] env[63531]: DEBUG nova.compute.manager [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 738.310110] env[63531]: DEBUG oslo_vmware.api [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Task: {'id': task-1117624, 'name': ReconfigVM_Task, 'duration_secs': 0.127918} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.310975] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244601', 'volume_id': '2a24c282-f681-40bf-a19c-5c711d342687', 'name': 'volume-2a24c282-f681-40bf-a19c-5c711d342687', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '410ddf7a-e45b-4769-a3db-5363ce2096a7', 'attached_at': '', 'detached_at': '', 'volume_id': '2a24c282-f681-40bf-a19c-5c711d342687', 'serial': '2a24c282-f681-40bf-a19c-5c711d342687'} {{(pid=63531) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 738.311523] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6ecac92c-42f8-4545-8b82-d31441b9acd2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.322673] env[63531]: DEBUG oslo_vmware.api [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Waiting for the task: (returnval){ [ 738.322673] env[63531]: value = "task-1117628" [ 738.322673] env[63531]: _type = "Task" [ 738.322673] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.329295] env[63531]: DEBUG nova.virt.hardware [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 738.329523] env[63531]: DEBUG nova.virt.hardware [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 738.329677] env[63531]: DEBUG nova.virt.hardware [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 738.329858] env[63531]: DEBUG nova.virt.hardware [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 738.330058] env[63531]: DEBUG nova.virt.hardware [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 738.330162] env[63531]: DEBUG nova.virt.hardware [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 738.330368] env[63531]: DEBUG nova.virt.hardware [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 738.330521] env[63531]: DEBUG nova.virt.hardware [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 738.330688] env[63531]: DEBUG nova.virt.hardware [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 738.330860] env[63531]: DEBUG nova.virt.hardware [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 738.331220] env[63531]: DEBUG nova.virt.hardware [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 738.332059] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f737a33-78ce-4441-82c5-3cb8cdd61783 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.338801] env[63531]: DEBUG oslo_vmware.api [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Task: {'id': task-1117628, 'name': Rename_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.343859] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af17f7a6-067d-4dd4-b676-8a027360a4d4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.360530] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e93d7fe6-1de5-4a53-ae44-efc3f5fc90f6 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Lock "5e9042b3-4dc2-4fa3-a664-c4b49a22e400" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 130.406s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 738.450394] env[63531]: DEBUG oslo_vmware.api [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Task: {'id': task-1117623, 'name': ReconfigVM_Task, 'duration_secs': 0.70318} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.450991] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Reconfigured VM instance instance-00000028 to attach disk [datastore1] c3920e60-b27f-496b-b616-afb23ab0bc26/c3920e60-b27f-496b-b616-afb23ab0bc26.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 738.453019] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3619cd15-a011-4d11-b0ee-f01be272ba28 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.459621] env[63531]: DEBUG oslo_vmware.api [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Waiting for the task: (returnval){ [ 738.459621] env[63531]: value = "task-1117629" [ 738.459621] env[63531]: _type = "Task" [ 738.459621] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.473624] env[63531]: DEBUG oslo_vmware.api [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Task: {'id': task-1117629, 'name': Rename_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.489684] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117627, 'name': CreateVM_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.655068] env[63531]: DEBUG nova.network.neutron [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Successfully updated port: 6102475b-8e43-4b77-85c3-5b574383c06d {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 738.759017] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.491s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 738.759017] env[63531]: DEBUG nova.compute.manager [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] [instance: 0af99240-fa7a-4eed-a729-a5ae98c41cf9] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 738.760820] env[63531]: DEBUG oslo_concurrency.lockutils [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 13.107s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 738.761181] env[63531]: DEBUG nova.objects.instance [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63531) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 738.800032] env[63531]: DEBUG oslo_concurrency.lockutils [req-77854249-de53-4228-aaf9-fe40b7eabc74 req-4476da96-3d24-45c3-a174-e5f7fc0f98d4 service nova] Releasing lock "refresh_cache-1ca7b851-2a06-4181-8271-58aafcd322d6" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 738.830969] env[63531]: DEBUG oslo_vmware.api [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Task: {'id': task-1117628, 'name': Rename_Task, 'duration_secs': 0.133753} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.831290] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 738.831531] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2f4de89b-7621-4943-955b-a33b4ea5fbe9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.838396] env[63531]: DEBUG oslo_vmware.api [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Waiting for the task: (returnval){ [ 738.838396] env[63531]: value = "task-1117630" [ 738.838396] env[63531]: _type = "Task" [ 738.838396] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.847966] env[63531]: DEBUG oslo_vmware.api [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Task: {'id': task-1117630, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.863817] env[63531]: DEBUG nova.compute.manager [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 738.970390] env[63531]: DEBUG oslo_vmware.api [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Task: {'id': task-1117629, 'name': Rename_Task, 'duration_secs': 0.295329} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.970664] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 738.970913] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-107608ba-9ece-4080-8e8e-713314ecde71 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.976842] env[63531]: DEBUG oslo_vmware.api [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Waiting for the task: (returnval){ [ 738.976842] env[63531]: value = "task-1117631" [ 738.976842] env[63531]: _type = "Task" [ 738.976842] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.988144] env[63531]: DEBUG oslo_vmware.api [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Task: {'id': task-1117631, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.991139] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117627, 'name': CreateVM_Task, 'duration_secs': 0.551372} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.991304] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 738.991957] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.992130] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.992453] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 738.992698] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d0c43c38-2bb7-4b66-8994-db30d65d6c5e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.997187] env[63531]: DEBUG oslo_vmware.api [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Waiting for the task: (returnval){ [ 738.997187] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52a3706a-25d7-57f5-1374-c99b560e149f" [ 738.997187] env[63531]: _type = "Task" [ 738.997187] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.004809] env[63531]: DEBUG oslo_vmware.api [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52a3706a-25d7-57f5-1374-c99b560e149f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.159353] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Acquiring lock "refresh_cache-e7fb00a3-0d2e-4f54-950d-337307112d7a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 739.159353] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Acquired lock "refresh_cache-e7fb00a3-0d2e-4f54-950d-337307112d7a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.159353] env[63531]: DEBUG nova.network.neutron [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 739.268823] env[63531]: DEBUG nova.compute.utils [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 739.270434] env[63531]: DEBUG nova.compute.manager [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] [instance: 0af99240-fa7a-4eed-a729-a5ae98c41cf9] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 739.270613] env[63531]: DEBUG nova.network.neutron [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] [instance: 0af99240-fa7a-4eed-a729-a5ae98c41cf9] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 739.334490] env[63531]: DEBUG nova.policy [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fc2f11ff65e943f9ba0feddfdee9596b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '67a25fa907944ea8aa86e3c953bfbc12', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 739.350069] env[63531]: DEBUG oslo_vmware.api [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Task: {'id': task-1117630, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.388682] env[63531]: DEBUG oslo_concurrency.lockutils [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 739.488078] env[63531]: DEBUG oslo_vmware.api [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Task: {'id': task-1117631, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.509158] env[63531]: DEBUG oslo_vmware.api [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52a3706a-25d7-57f5-1374-c99b560e149f, 'name': SearchDatastore_Task, 'duration_secs': 0.015055} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.509319] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 739.509567] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 739.509809] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 739.509957] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.510157] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 739.510432] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f0d54a00-9457-4a80-8228-2a56cf2c4200 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.521026] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 739.521233] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 739.522245] env[63531]: DEBUG nova.compute.manager [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Stashing vm_state: active {{(pid=63531) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 739.525076] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ca9a29a6-cc6c-41af-89c7-9ec68797a35c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.532480] env[63531]: DEBUG oslo_vmware.api [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Waiting for the task: (returnval){ [ 739.532480] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52def764-acfb-a8cf-17c1-64369d3530ac" [ 739.532480] env[63531]: _type = "Task" [ 739.532480] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.542220] env[63531]: DEBUG oslo_vmware.api [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52def764-acfb-a8cf-17c1-64369d3530ac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.685376] env[63531]: DEBUG nova.network.neutron [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] [instance: 0af99240-fa7a-4eed-a729-a5ae98c41cf9] Successfully created port: c2356641-c89b-4610-ade8-8cc4e392021b {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 739.711377] env[63531]: DEBUG nova.network.neutron [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 739.771874] env[63531]: DEBUG oslo_concurrency.lockutils [None req-140b7982-49e8-446b-b5f5-06f706376269 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.011s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 739.773154] env[63531]: DEBUG oslo_concurrency.lockutils [None req-396bcb9b-cfea-4ccb-9d40-6e75c674fc51 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.479s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 739.777021] env[63531]: DEBUG nova.objects.instance [None req-396bcb9b-cfea-4ccb-9d40-6e75c674fc51 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Lazy-loading 'resources' on Instance uuid ffb34f90-d22f-440b-ba07-75d474c3c300 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 739.777021] env[63531]: DEBUG nova.compute.manager [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] [instance: 0af99240-fa7a-4eed-a729-a5ae98c41cf9] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 739.787360] env[63531]: DEBUG nova.compute.manager [req-e58fa8cb-8ede-4e23-8841-52fd5269bff7 req-14ede377-9b44-4880-a49f-74d8419d8064 service nova] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Received event network-vif-plugged-6102475b-8e43-4b77-85c3-5b574383c06d {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 739.787404] env[63531]: DEBUG oslo_concurrency.lockutils [req-e58fa8cb-8ede-4e23-8841-52fd5269bff7 req-14ede377-9b44-4880-a49f-74d8419d8064 service nova] Acquiring lock "e7fb00a3-0d2e-4f54-950d-337307112d7a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 739.788672] env[63531]: DEBUG oslo_concurrency.lockutils [req-e58fa8cb-8ede-4e23-8841-52fd5269bff7 req-14ede377-9b44-4880-a49f-74d8419d8064 service nova] Lock "e7fb00a3-0d2e-4f54-950d-337307112d7a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 739.788672] env[63531]: DEBUG oslo_concurrency.lockutils [req-e58fa8cb-8ede-4e23-8841-52fd5269bff7 req-14ede377-9b44-4880-a49f-74d8419d8064 service nova] Lock "e7fb00a3-0d2e-4f54-950d-337307112d7a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 739.788672] env[63531]: DEBUG nova.compute.manager [req-e58fa8cb-8ede-4e23-8841-52fd5269bff7 req-14ede377-9b44-4880-a49f-74d8419d8064 service nova] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] No waiting events found dispatching network-vif-plugged-6102475b-8e43-4b77-85c3-5b574383c06d {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 739.788672] env[63531]: WARNING nova.compute.manager [req-e58fa8cb-8ede-4e23-8841-52fd5269bff7 req-14ede377-9b44-4880-a49f-74d8419d8064 service nova] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Received unexpected event network-vif-plugged-6102475b-8e43-4b77-85c3-5b574383c06d for instance with vm_state building and task_state spawning. [ 739.788672] env[63531]: DEBUG nova.compute.manager [req-e58fa8cb-8ede-4e23-8841-52fd5269bff7 req-14ede377-9b44-4880-a49f-74d8419d8064 service nova] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Received event network-changed-6102475b-8e43-4b77-85c3-5b574383c06d {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 739.788869] env[63531]: DEBUG nova.compute.manager [req-e58fa8cb-8ede-4e23-8841-52fd5269bff7 req-14ede377-9b44-4880-a49f-74d8419d8064 service nova] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Refreshing instance network info cache due to event network-changed-6102475b-8e43-4b77-85c3-5b574383c06d. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 739.788869] env[63531]: DEBUG oslo_concurrency.lockutils [req-e58fa8cb-8ede-4e23-8841-52fd5269bff7 req-14ede377-9b44-4880-a49f-74d8419d8064 service nova] Acquiring lock "refresh_cache-e7fb00a3-0d2e-4f54-950d-337307112d7a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 739.852022] env[63531]: DEBUG oslo_vmware.api [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Task: {'id': task-1117630, 'name': PowerOnVM_Task, 'duration_secs': 0.671} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.852022] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 739.852022] env[63531]: INFO nova.compute.manager [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Took 4.84 seconds to spawn the instance on the hypervisor. [ 739.852022] env[63531]: DEBUG nova.compute.manager [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 739.852760] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d414fb8-a3ca-463e-8beb-4dd0c6cdb830 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.991228] env[63531]: DEBUG oslo_vmware.api [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Task: {'id': task-1117631, 'name': PowerOnVM_Task, 'duration_secs': 0.565885} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.991685] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 739.993824] env[63531]: INFO nova.compute.manager [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Took 9.19 seconds to spawn the instance on the hypervisor. [ 739.994762] env[63531]: DEBUG nova.compute.manager [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 739.996677] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-862429de-604e-4887-bc46-a1e6a5d6c979 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.044759] env[63531]: DEBUG oslo_vmware.api [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52def764-acfb-a8cf-17c1-64369d3530ac, 'name': SearchDatastore_Task, 'duration_secs': 0.01178} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.044759] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-94f95517-f570-49c4-ac5b-19da6763c625 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.050722] env[63531]: DEBUG oslo_vmware.api [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Waiting for the task: (returnval){ [ 740.050722] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5261052c-ce33-3526-bbf3-b30bacca3dc8" [ 740.050722] env[63531]: _type = "Task" [ 740.050722] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.061619] env[63531]: DEBUG oslo_vmware.api [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5261052c-ce33-3526-bbf3-b30bacca3dc8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.080449] env[63531]: DEBUG oslo_concurrency.lockutils [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 740.172669] env[63531]: DEBUG nova.network.neutron [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Updating instance_info_cache with network_info: [{"id": "6102475b-8e43-4b77-85c3-5b574383c06d", "address": "fa:16:3e:9c:74:35", "network": {"id": "dc416058-1c04-47ac-8b73-1557f7c33d1c", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1350689703-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "824ed405cc6e4868a70e82489e05e8ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "05b1253d-2b87-4158-9ff1-dafcf829f11f", "external-id": "nsx-vlan-transportzone-55", "segmentation_id": 55, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6102475b-8e", "ovs_interfaceid": "6102475b-8e43-4b77-85c3-5b574383c06d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.370605] env[63531]: INFO nova.compute.manager [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Took 24.62 seconds to build instance. [ 740.527484] env[63531]: INFO nova.compute.manager [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Took 26.01 seconds to build instance. [ 740.564299] env[63531]: DEBUG oslo_vmware.api [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5261052c-ce33-3526-bbf3-b30bacca3dc8, 'name': SearchDatastore_Task, 'duration_secs': 0.009936} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.564571] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.564864] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] 1ca7b851-2a06-4181-8271-58aafcd322d6/1ca7b851-2a06-4181-8271-58aafcd322d6.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 740.565153] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8175db6e-7fcf-40b2-815d-f29ae3edaebd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.572841] env[63531]: DEBUG oslo_vmware.api [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Waiting for the task: (returnval){ [ 740.572841] env[63531]: value = "task-1117632" [ 740.572841] env[63531]: _type = "Task" [ 740.572841] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.583347] env[63531]: DEBUG oslo_vmware.api [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1117632, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.669136] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ce9fc42-57ca-4470-ab3c-d392bbb298d4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.676078] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Releasing lock "refresh_cache-e7fb00a3-0d2e-4f54-950d-337307112d7a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.676877] env[63531]: DEBUG nova.compute.manager [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Instance network_info: |[{"id": "6102475b-8e43-4b77-85c3-5b574383c06d", "address": "fa:16:3e:9c:74:35", "network": {"id": "dc416058-1c04-47ac-8b73-1557f7c33d1c", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1350689703-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "824ed405cc6e4868a70e82489e05e8ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "05b1253d-2b87-4158-9ff1-dafcf829f11f", "external-id": "nsx-vlan-transportzone-55", "segmentation_id": 55, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6102475b-8e", "ovs_interfaceid": "6102475b-8e43-4b77-85c3-5b574383c06d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 740.682063] env[63531]: DEBUG oslo_concurrency.lockutils [req-e58fa8cb-8ede-4e23-8841-52fd5269bff7 req-14ede377-9b44-4880-a49f-74d8419d8064 service nova] Acquired lock "refresh_cache-e7fb00a3-0d2e-4f54-950d-337307112d7a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.682063] env[63531]: DEBUG nova.network.neutron [req-e58fa8cb-8ede-4e23-8841-52fd5269bff7 req-14ede377-9b44-4880-a49f-74d8419d8064 service nova] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Refreshing network info cache for port 6102475b-8e43-4b77-85c3-5b574383c06d {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 740.683664] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9c:74:35', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '05b1253d-2b87-4158-9ff1-dafcf829f11f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6102475b-8e43-4b77-85c3-5b574383c06d', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 740.697226] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Creating folder: Project (824ed405cc6e4868a70e82489e05e8ac). Parent ref: group-v244585. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 740.698573] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a7b37db4-90e6-4891-abb5-86b78646156a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.701119] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a26bfaf-f13d-47ab-bfe9-34c4d592ba37 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.737751] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71c3b6cc-58be-463c-9218-d2de7d9d67db {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.739846] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Created folder: Project (824ed405cc6e4868a70e82489e05e8ac) in parent group-v244585. [ 740.739846] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Creating folder: Instances. Parent ref: group-v244631. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 740.740155] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ad1aed63-f68f-4427-a1e8-bc5039a94b36 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.748088] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1e68750-923a-4621-a650-47dddd1b4b0d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.752810] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Created folder: Instances in parent group-v244631. [ 740.753070] env[63531]: DEBUG oslo.service.loopingcall [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 740.753605] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 740.753806] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dfa60572-3265-4256-b97d-d4765f77fb8e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.777458] env[63531]: DEBUG nova.compute.provider_tree [None req-396bcb9b-cfea-4ccb-9d40-6e75c674fc51 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 740.783373] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 740.783373] env[63531]: value = "task-1117635" [ 740.783373] env[63531]: _type = "Task" [ 740.783373] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.787427] env[63531]: DEBUG nova.compute.manager [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] [instance: 0af99240-fa7a-4eed-a729-a5ae98c41cf9] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 740.795439] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117635, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.816202] env[63531]: DEBUG nova.virt.hardware [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 740.816485] env[63531]: DEBUG nova.virt.hardware [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 740.816679] env[63531]: DEBUG nova.virt.hardware [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 740.816884] env[63531]: DEBUG nova.virt.hardware [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 740.817055] env[63531]: DEBUG nova.virt.hardware [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 740.817242] env[63531]: DEBUG nova.virt.hardware [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 740.817506] env[63531]: DEBUG nova.virt.hardware [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 740.817731] env[63531]: DEBUG nova.virt.hardware [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 740.818073] env[63531]: DEBUG nova.virt.hardware [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 740.818291] env[63531]: DEBUG nova.virt.hardware [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 740.818508] env[63531]: DEBUG nova.virt.hardware [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 740.819520] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-463c034e-b513-450f-93b6-ff4ca1f07430 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.831025] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a75e672-fc95-4fc2-8501-b61f60e158a0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.872439] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b3bf014d-e3c5-4e97-8f4f-b18f657f8c9a tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Lock "410ddf7a-e45b-4769-a3db-5363ce2096a7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 125.455s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.032304] env[63531]: DEBUG oslo_concurrency.lockutils [None req-920a84d5-681e-4ac6-bc87-f69a9583ef48 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Lock "c3920e60-b27f-496b-b616-afb23ab0bc26" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 130.471s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.042540] env[63531]: INFO nova.compute.manager [None req-75b88ae3-9411-41cb-88e2-c8261dabea8d tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Rescuing [ 741.042845] env[63531]: DEBUG oslo_concurrency.lockutils [None req-75b88ae3-9411-41cb-88e2-c8261dabea8d tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Acquiring lock "refresh_cache-c3920e60-b27f-496b-b616-afb23ab0bc26" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 741.043998] env[63531]: DEBUG oslo_concurrency.lockutils [None req-75b88ae3-9411-41cb-88e2-c8261dabea8d tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Acquired lock "refresh_cache-c3920e60-b27f-496b-b616-afb23ab0bc26" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 741.043998] env[63531]: DEBUG nova.network.neutron [None req-75b88ae3-9411-41cb-88e2-c8261dabea8d tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 741.084966] env[63531]: DEBUG oslo_vmware.api [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1117632, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.283288] env[63531]: DEBUG nova.scheduler.client.report [None req-396bcb9b-cfea-4ccb-9d40-6e75c674fc51 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 741.297566] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117635, 'name': CreateVM_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.317376] env[63531]: DEBUG nova.network.neutron [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] [instance: 0af99240-fa7a-4eed-a729-a5ae98c41cf9] Successfully updated port: c2356641-c89b-4610-ade8-8cc4e392021b {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 741.376650] env[63531]: DEBUG nova.compute.manager [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 741.535287] env[63531]: DEBUG nova.compute.manager [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 741.586805] env[63531]: DEBUG oslo_vmware.api [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1117632, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.698727} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.586985] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] 1ca7b851-2a06-4181-8271-58aafcd322d6/1ca7b851-2a06-4181-8271-58aafcd322d6.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 741.587218] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 741.589026] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2a74266f-23b5-454c-9bc7-011098530c60 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.595484] env[63531]: DEBUG oslo_vmware.api [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Waiting for the task: (returnval){ [ 741.595484] env[63531]: value = "task-1117636" [ 741.595484] env[63531]: _type = "Task" [ 741.595484] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.604433] env[63531]: DEBUG oslo_vmware.api [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1117636, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.732512] env[63531]: DEBUG nova.network.neutron [None req-75b88ae3-9411-41cb-88e2-c8261dabea8d tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Updating instance_info_cache with network_info: [{"id": "91733d5a-84d6-48df-ba4d-7e362a642282", "address": "fa:16:3e:df:a1:31", "network": {"id": "d28b2087-329c-4917-aa6e-83ab83a63a74", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1054369059-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a634a61020b94d62a8fb37a3f96acc10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91733d5a-84", "ovs_interfaceid": "91733d5a-84d6-48df-ba4d-7e362a642282", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.792974] env[63531]: DEBUG oslo_concurrency.lockutils [None req-396bcb9b-cfea-4ccb-9d40-6e75c674fc51 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.020s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.798300] env[63531]: DEBUG oslo_concurrency.lockutils [None req-835ac754-8e34-4ef9-903d-b2d75006b444 tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.889s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 741.798527] env[63531]: DEBUG nova.objects.instance [None req-835ac754-8e34-4ef9-903d-b2d75006b444 tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Lazy-loading 'resources' on Instance uuid 7bc63fdd-4b50-486d-9681-ca0baa08f7b3 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 741.799686] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117635, 'name': CreateVM_Task, 'duration_secs': 0.528833} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.799782] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 741.800406] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 741.800562] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 741.800861] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 741.801114] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e41f9973-5493-420f-a7fb-a0d0b5a62987 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.812179] env[63531]: DEBUG oslo_vmware.api [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Waiting for the task: (returnval){ [ 741.812179] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52ad66bb-21ca-912a-82be-5ab60009d1f7" [ 741.812179] env[63531]: _type = "Task" [ 741.812179] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.824782] env[63531]: DEBUG oslo_vmware.api [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52ad66bb-21ca-912a-82be-5ab60009d1f7, 'name': SearchDatastore_Task, 'duration_secs': 0.009193} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.824924] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 741.825178] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 741.825418] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 741.825737] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 741.825894] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 741.831510] env[63531]: INFO nova.scheduler.client.report [None req-396bcb9b-cfea-4ccb-9d40-6e75c674fc51 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Deleted allocations for instance ffb34f90-d22f-440b-ba07-75d474c3c300 [ 741.832667] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Acquiring lock "refresh_cache-0af99240-fa7a-4eed-a729-a5ae98c41cf9" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 741.832803] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Acquired lock "refresh_cache-0af99240-fa7a-4eed-a729-a5ae98c41cf9" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 741.832943] env[63531]: DEBUG nova.network.neutron [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] [instance: 0af99240-fa7a-4eed-a729-a5ae98c41cf9] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 741.833941] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-54f24af9-6560-4d5d-90a9-e38fb275c1d1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.848445] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 741.848645] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 741.849482] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-56d50050-edd8-4191-8b80-48cc799713d6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.852572] env[63531]: DEBUG nova.network.neutron [req-e58fa8cb-8ede-4e23-8841-52fd5269bff7 req-14ede377-9b44-4880-a49f-74d8419d8064 service nova] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Updated VIF entry in instance network info cache for port 6102475b-8e43-4b77-85c3-5b574383c06d. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 741.852953] env[63531]: DEBUG nova.network.neutron [req-e58fa8cb-8ede-4e23-8841-52fd5269bff7 req-14ede377-9b44-4880-a49f-74d8419d8064 service nova] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Updating instance_info_cache with network_info: [{"id": "6102475b-8e43-4b77-85c3-5b574383c06d", "address": "fa:16:3e:9c:74:35", "network": {"id": "dc416058-1c04-47ac-8b73-1557f7c33d1c", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1350689703-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "824ed405cc6e4868a70e82489e05e8ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "05b1253d-2b87-4158-9ff1-dafcf829f11f", "external-id": "nsx-vlan-transportzone-55", "segmentation_id": 55, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6102475b-8e", "ovs_interfaceid": "6102475b-8e43-4b77-85c3-5b574383c06d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.858318] env[63531]: DEBUG oslo_vmware.api [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Waiting for the task: (returnval){ [ 741.858318] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]529ab4f9-2cbc-3dc6-7f7a-60ccc1d12e5c" [ 741.858318] env[63531]: _type = "Task" [ 741.858318] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.874072] env[63531]: DEBUG oslo_vmware.api [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]529ab4f9-2cbc-3dc6-7f7a-60ccc1d12e5c, 'name': SearchDatastore_Task, 'duration_secs': 0.009902} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.874889] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-55397eee-9624-4ab2-923a-2749136d6eec {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.880660] env[63531]: DEBUG oslo_vmware.api [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Waiting for the task: (returnval){ [ 741.880660] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5219adb9-32b2-5aa6-556e-3086298f7351" [ 741.880660] env[63531]: _type = "Task" [ 741.880660] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.894519] env[63531]: DEBUG oslo_vmware.api [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5219adb9-32b2-5aa6-556e-3086298f7351, 'name': SearchDatastore_Task, 'duration_secs': 0.009783} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.894519] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 741.894519] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] e7fb00a3-0d2e-4f54-950d-337307112d7a/e7fb00a3-0d2e-4f54-950d-337307112d7a.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 741.894772] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cc115e3c-cbe2-432d-934c-6e25bf8fd426 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.903125] env[63531]: DEBUG oslo_vmware.api [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Waiting for the task: (returnval){ [ 741.903125] env[63531]: value = "task-1117637" [ 741.903125] env[63531]: _type = "Task" [ 741.903125] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.907529] env[63531]: DEBUG oslo_concurrency.lockutils [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 741.910761] env[63531]: DEBUG oslo_vmware.api [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117637, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.931508] env[63531]: DEBUG nova.compute.manager [req-a3ed3488-ac22-4945-8c1a-95333dc73f3a req-1005fc31-4525-454d-a616-36ec616d6e6d service nova] [instance: 0af99240-fa7a-4eed-a729-a5ae98c41cf9] Received event network-vif-plugged-c2356641-c89b-4610-ade8-8cc4e392021b {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 741.931508] env[63531]: DEBUG oslo_concurrency.lockutils [req-a3ed3488-ac22-4945-8c1a-95333dc73f3a req-1005fc31-4525-454d-a616-36ec616d6e6d service nova] Acquiring lock "0af99240-fa7a-4eed-a729-a5ae98c41cf9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 741.931666] env[63531]: DEBUG oslo_concurrency.lockutils [req-a3ed3488-ac22-4945-8c1a-95333dc73f3a req-1005fc31-4525-454d-a616-36ec616d6e6d service nova] Lock "0af99240-fa7a-4eed-a729-a5ae98c41cf9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 741.931860] env[63531]: DEBUG oslo_concurrency.lockutils [req-a3ed3488-ac22-4945-8c1a-95333dc73f3a req-1005fc31-4525-454d-a616-36ec616d6e6d service nova] Lock "0af99240-fa7a-4eed-a729-a5ae98c41cf9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.931938] env[63531]: DEBUG nova.compute.manager [req-a3ed3488-ac22-4945-8c1a-95333dc73f3a req-1005fc31-4525-454d-a616-36ec616d6e6d service nova] [instance: 0af99240-fa7a-4eed-a729-a5ae98c41cf9] No waiting events found dispatching network-vif-plugged-c2356641-c89b-4610-ade8-8cc4e392021b {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 741.932086] env[63531]: WARNING nova.compute.manager [req-a3ed3488-ac22-4945-8c1a-95333dc73f3a req-1005fc31-4525-454d-a616-36ec616d6e6d service nova] [instance: 0af99240-fa7a-4eed-a729-a5ae98c41cf9] Received unexpected event network-vif-plugged-c2356641-c89b-4610-ade8-8cc4e392021b for instance with vm_state building and task_state spawning. [ 741.932253] env[63531]: DEBUG nova.compute.manager [req-a3ed3488-ac22-4945-8c1a-95333dc73f3a req-1005fc31-4525-454d-a616-36ec616d6e6d service nova] [instance: 0af99240-fa7a-4eed-a729-a5ae98c41cf9] Received event network-changed-c2356641-c89b-4610-ade8-8cc4e392021b {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 741.932407] env[63531]: DEBUG nova.compute.manager [req-a3ed3488-ac22-4945-8c1a-95333dc73f3a req-1005fc31-4525-454d-a616-36ec616d6e6d service nova] [instance: 0af99240-fa7a-4eed-a729-a5ae98c41cf9] Refreshing instance network info cache due to event network-changed-c2356641-c89b-4610-ade8-8cc4e392021b. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 741.932571] env[63531]: DEBUG oslo_concurrency.lockutils [req-a3ed3488-ac22-4945-8c1a-95333dc73f3a req-1005fc31-4525-454d-a616-36ec616d6e6d service nova] Acquiring lock "refresh_cache-0af99240-fa7a-4eed-a729-a5ae98c41cf9" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 742.054731] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 742.109903] env[63531]: DEBUG oslo_vmware.api [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1117636, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.100888} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.110142] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 742.110956] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-645a5ab9-bb62-45dc-9366-3b14ab7ca4b3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.135737] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Reconfiguring VM instance instance-0000002a to attach disk [datastore2] 1ca7b851-2a06-4181-8271-58aafcd322d6/1ca7b851-2a06-4181-8271-58aafcd322d6.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 742.136425] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4f97cb4c-031f-45bf-86fc-fd7c34ce5fb9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.158463] env[63531]: DEBUG oslo_vmware.api [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Waiting for the task: (returnval){ [ 742.158463] env[63531]: value = "task-1117638" [ 742.158463] env[63531]: _type = "Task" [ 742.158463] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.168929] env[63531]: DEBUG oslo_vmware.api [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1117638, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.236431] env[63531]: DEBUG oslo_concurrency.lockutils [None req-75b88ae3-9411-41cb-88e2-c8261dabea8d tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Releasing lock "refresh_cache-c3920e60-b27f-496b-b616-afb23ab0bc26" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.347775] env[63531]: DEBUG oslo_concurrency.lockutils [None req-396bcb9b-cfea-4ccb-9d40-6e75c674fc51 tempest-ServersAdminTestJSON-195653164 tempest-ServersAdminTestJSON-195653164-project-member] Lock "ffb34f90-d22f-440b-ba07-75d474c3c300" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.784s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 742.355443] env[63531]: DEBUG oslo_concurrency.lockutils [req-e58fa8cb-8ede-4e23-8841-52fd5269bff7 req-14ede377-9b44-4880-a49f-74d8419d8064 service nova] Releasing lock "refresh_cache-e7fb00a3-0d2e-4f54-950d-337307112d7a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.420738] env[63531]: DEBUG oslo_vmware.api [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117637, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.421627] env[63531]: DEBUG nova.network.neutron [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] [instance: 0af99240-fa7a-4eed-a729-a5ae98c41cf9] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 742.667694] env[63531]: DEBUG oslo_vmware.api [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1117638, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.684573] env[63531]: DEBUG nova.network.neutron [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] [instance: 0af99240-fa7a-4eed-a729-a5ae98c41cf9] Updating instance_info_cache with network_info: [{"id": "c2356641-c89b-4610-ade8-8cc4e392021b", "address": "fa:16:3e:46:eb:a0", "network": {"id": "5439232a-ec6d-4184-a9f1-f7be5c1526bc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.186", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "54fbef8a6ee0488f820706d2036e9c99", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f267bcdd-0daa-4337-9709-5fc060c267d8", "external-id": "nsx-vlan-transportzone-308", "segmentation_id": 308, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2356641-c8", "ovs_interfaceid": "c2356641-c89b-4610-ade8-8cc4e392021b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.751007] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05427f08-3d46-4e31-a8f0-9bc4419bf642 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.760045] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f21f9e6b-9aa7-48c1-a138-5695f2dd098a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.800379] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-75b88ae3-9411-41cb-88e2-c8261dabea8d tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 742.801448] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff575cee-3025-475d-a727-925285fe33f6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.804440] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-79e11967-ca75-4ac8-976c-d827b87b147c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.813246] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a73dd6b-6842-468c-9560-73b51dff1103 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.817454] env[63531]: DEBUG oslo_vmware.api [None req-75b88ae3-9411-41cb-88e2-c8261dabea8d tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Waiting for the task: (returnval){ [ 742.817454] env[63531]: value = "task-1117639" [ 742.817454] env[63531]: _type = "Task" [ 742.817454] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.829775] env[63531]: DEBUG nova.compute.provider_tree [None req-835ac754-8e34-4ef9-903d-b2d75006b444 tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 742.836851] env[63531]: DEBUG oslo_vmware.api [None req-75b88ae3-9411-41cb-88e2-c8261dabea8d tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Task: {'id': task-1117639, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.915286] env[63531]: DEBUG oslo_vmware.api [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117637, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.933472] env[63531]: DEBUG nova.compute.manager [req-dcce3a33-952e-4d14-9782-4542fbc125a8 req-a729f9d1-7965-4630-864b-f8a3a204d939 service nova] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Received event network-changed-c173088d-cfd8-4f05-8936-1d0d5927ae34 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 742.934026] env[63531]: DEBUG nova.compute.manager [req-dcce3a33-952e-4d14-9782-4542fbc125a8 req-a729f9d1-7965-4630-864b-f8a3a204d939 service nova] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Refreshing instance network info cache due to event network-changed-c173088d-cfd8-4f05-8936-1d0d5927ae34. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 742.934390] env[63531]: DEBUG oslo_concurrency.lockutils [req-dcce3a33-952e-4d14-9782-4542fbc125a8 req-a729f9d1-7965-4630-864b-f8a3a204d939 service nova] Acquiring lock "refresh_cache-410ddf7a-e45b-4769-a3db-5363ce2096a7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 742.934697] env[63531]: DEBUG oslo_concurrency.lockutils [req-dcce3a33-952e-4d14-9782-4542fbc125a8 req-a729f9d1-7965-4630-864b-f8a3a204d939 service nova] Acquired lock "refresh_cache-410ddf7a-e45b-4769-a3db-5363ce2096a7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.934957] env[63531]: DEBUG nova.network.neutron [req-dcce3a33-952e-4d14-9782-4542fbc125a8 req-a729f9d1-7965-4630-864b-f8a3a204d939 service nova] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Refreshing network info cache for port c173088d-cfd8-4f05-8936-1d0d5927ae34 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 743.168156] env[63531]: DEBUG oslo_vmware.api [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1117638, 'name': ReconfigVM_Task, 'duration_secs': 0.728023} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.168459] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Reconfigured VM instance instance-0000002a to attach disk [datastore2] 1ca7b851-2a06-4181-8271-58aafcd322d6/1ca7b851-2a06-4181-8271-58aafcd322d6.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 743.169091] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-36fbe4b8-c0db-43b2-8763-c9c416fafb99 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.175433] env[63531]: DEBUG oslo_vmware.api [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Waiting for the task: (returnval){ [ 743.175433] env[63531]: value = "task-1117640" [ 743.175433] env[63531]: _type = "Task" [ 743.175433] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.182993] env[63531]: DEBUG oslo_vmware.api [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1117640, 'name': Rename_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.192790] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Releasing lock "refresh_cache-0af99240-fa7a-4eed-a729-a5ae98c41cf9" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 743.192888] env[63531]: DEBUG nova.compute.manager [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] [instance: 0af99240-fa7a-4eed-a729-a5ae98c41cf9] Instance network_info: |[{"id": "c2356641-c89b-4610-ade8-8cc4e392021b", "address": "fa:16:3e:46:eb:a0", "network": {"id": "5439232a-ec6d-4184-a9f1-f7be5c1526bc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.186", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "54fbef8a6ee0488f820706d2036e9c99", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f267bcdd-0daa-4337-9709-5fc060c267d8", "external-id": "nsx-vlan-transportzone-308", "segmentation_id": 308, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2356641-c8", "ovs_interfaceid": "c2356641-c89b-4610-ade8-8cc4e392021b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 743.193147] env[63531]: DEBUG oslo_concurrency.lockutils [req-a3ed3488-ac22-4945-8c1a-95333dc73f3a req-1005fc31-4525-454d-a616-36ec616d6e6d service nova] Acquired lock "refresh_cache-0af99240-fa7a-4eed-a729-a5ae98c41cf9" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.193339] env[63531]: DEBUG nova.network.neutron [req-a3ed3488-ac22-4945-8c1a-95333dc73f3a req-1005fc31-4525-454d-a616-36ec616d6e6d service nova] [instance: 0af99240-fa7a-4eed-a729-a5ae98c41cf9] Refreshing network info cache for port c2356641-c89b-4610-ade8-8cc4e392021b {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 743.198022] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] [instance: 0af99240-fa7a-4eed-a729-a5ae98c41cf9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:46:eb:a0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f267bcdd-0daa-4337-9709-5fc060c267d8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c2356641-c89b-4610-ade8-8cc4e392021b', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 743.202129] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Creating folder: Project (67a25fa907944ea8aa86e3c953bfbc12). Parent ref: group-v244585. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 743.202931] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5d520fb2-0791-42d3-9d2e-af1e97cc7728 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.214836] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Created folder: Project (67a25fa907944ea8aa86e3c953bfbc12) in parent group-v244585. [ 743.215084] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Creating folder: Instances. Parent ref: group-v244634. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 743.215333] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3a9aced6-daba-4806-933d-3d32b6330ac5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.224563] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Created folder: Instances in parent group-v244634. [ 743.224851] env[63531]: DEBUG oslo.service.loopingcall [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 743.225090] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0af99240-fa7a-4eed-a729-a5ae98c41cf9] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 743.225343] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e2ef36dc-1488-49a2-a57d-72f91b231fbc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.248285] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 743.248285] env[63531]: value = "task-1117643" [ 743.248285] env[63531]: _type = "Task" [ 743.248285] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.256351] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117643, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.327160] env[63531]: DEBUG oslo_vmware.api [None req-75b88ae3-9411-41cb-88e2-c8261dabea8d tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Task: {'id': task-1117639, 'name': PowerOffVM_Task, 'duration_secs': 0.178634} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.327437] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-75b88ae3-9411-41cb-88e2-c8261dabea8d tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 743.328255] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45608159-a538-46f4-8ddf-74b7bdf4d22d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.332696] env[63531]: DEBUG nova.scheduler.client.report [None req-835ac754-8e34-4ef9-903d-b2d75006b444 tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 743.350778] env[63531]: DEBUG oslo_concurrency.lockutils [None req-835ac754-8e34-4ef9-903d-b2d75006b444 tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.552s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 743.353349] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a4c7c9f-33e6-422d-b062-e4679fe1e1f7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.356270] env[63531]: DEBUG oslo_concurrency.lockutils [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.271s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 743.357818] env[63531]: INFO nova.compute.claims [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 743.384136] env[63531]: INFO nova.scheduler.client.report [None req-835ac754-8e34-4ef9-903d-b2d75006b444 tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Deleted allocations for instance 7bc63fdd-4b50-486d-9681-ca0baa08f7b3 [ 743.396774] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-75b88ae3-9411-41cb-88e2-c8261dabea8d tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 743.397375] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6c888554-3099-42b6-a5f2-cd9f58f4b9d6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.406396] env[63531]: DEBUG oslo_vmware.api [None req-75b88ae3-9411-41cb-88e2-c8261dabea8d tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Waiting for the task: (returnval){ [ 743.406396] env[63531]: value = "task-1117644" [ 743.406396] env[63531]: _type = "Task" [ 743.406396] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.420971] env[63531]: DEBUG oslo_vmware.api [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117637, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.422096] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-75b88ae3-9411-41cb-88e2-c8261dabea8d tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] VM already powered off {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 743.422735] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-75b88ae3-9411-41cb-88e2-c8261dabea8d tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 743.422735] env[63531]: DEBUG oslo_concurrency.lockutils [None req-75b88ae3-9411-41cb-88e2-c8261dabea8d tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 743.425566] env[63531]: DEBUG oslo_concurrency.lockutils [None req-75b88ae3-9411-41cb-88e2-c8261dabea8d tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.425566] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-75b88ae3-9411-41cb-88e2-c8261dabea8d tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 743.425566] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-62faec37-f0f3-41ac-849c-98449cb0bfc1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.431473] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-75b88ae3-9411-41cb-88e2-c8261dabea8d tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 743.431650] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-75b88ae3-9411-41cb-88e2-c8261dabea8d tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 743.432956] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f8cfbfd1-3b8d-4a89-b509-5dc523f1a4fb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.441442] env[63531]: DEBUG oslo_vmware.api [None req-75b88ae3-9411-41cb-88e2-c8261dabea8d tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Waiting for the task: (returnval){ [ 743.441442] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5235a915-4176-5e07-9fbe-ea06d002a860" [ 743.441442] env[63531]: _type = "Task" [ 743.441442] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.451884] env[63531]: DEBUG oslo_vmware.api [None req-75b88ae3-9411-41cb-88e2-c8261dabea8d tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5235a915-4176-5e07-9fbe-ea06d002a860, 'name': SearchDatastore_Task, 'duration_secs': 0.008966} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.452832] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-09c68de9-0d5a-4965-a5df-8463228b3def {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.458727] env[63531]: DEBUG oslo_vmware.api [None req-75b88ae3-9411-41cb-88e2-c8261dabea8d tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Waiting for the task: (returnval){ [ 743.458727] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52c4ff64-1a30-606d-be31-ccc7af768f29" [ 743.458727] env[63531]: _type = "Task" [ 743.458727] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.467651] env[63531]: DEBUG oslo_vmware.api [None req-75b88ae3-9411-41cb-88e2-c8261dabea8d tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52c4ff64-1a30-606d-be31-ccc7af768f29, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.687318] env[63531]: DEBUG oslo_vmware.api [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1117640, 'name': Rename_Task, 'duration_secs': 0.148239} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.687611] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 743.687856] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f0c3055b-3b36-4e83-81c7-0e5ac331ba3d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.694748] env[63531]: DEBUG oslo_vmware.api [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Waiting for the task: (returnval){ [ 743.694748] env[63531]: value = "task-1117645" [ 743.694748] env[63531]: _type = "Task" [ 743.694748] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.705730] env[63531]: DEBUG oslo_vmware.api [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1117645, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.761199] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117643, 'name': CreateVM_Task, 'duration_secs': 0.346115} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.766071] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0af99240-fa7a-4eed-a729-a5ae98c41cf9] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 743.769747] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 743.769747] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.769747] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 743.769747] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e39f006f-7b7a-4407-aa70-77dda3c241d5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.774463] env[63531]: DEBUG oslo_vmware.api [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Waiting for the task: (returnval){ [ 743.774463] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52dda030-d3b8-bc1e-b4b8-0fd4dc38877d" [ 743.774463] env[63531]: _type = "Task" [ 743.774463] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.782928] env[63531]: DEBUG oslo_vmware.api [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52dda030-d3b8-bc1e-b4b8-0fd4dc38877d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.807439] env[63531]: DEBUG nova.network.neutron [req-dcce3a33-952e-4d14-9782-4542fbc125a8 req-a729f9d1-7965-4630-864b-f8a3a204d939 service nova] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Updated VIF entry in instance network info cache for port c173088d-cfd8-4f05-8936-1d0d5927ae34. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 743.807824] env[63531]: DEBUG nova.network.neutron [req-dcce3a33-952e-4d14-9782-4542fbc125a8 req-a729f9d1-7965-4630-864b-f8a3a204d939 service nova] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Updating instance_info_cache with network_info: [{"id": "c173088d-cfd8-4f05-8936-1d0d5927ae34", "address": "fa:16:3e:24:0f:e5", "network": {"id": "88ee583d-32f8-4f4e-b02c-392260bab32b", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-2110978471-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.181", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fed6b4862d76469b87ecf16dc6dce177", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc173088d-cf", "ovs_interfaceid": "c173088d-cfd8-4f05-8936-1d0d5927ae34", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.896855] env[63531]: DEBUG oslo_concurrency.lockutils [None req-835ac754-8e34-4ef9-903d-b2d75006b444 tempest-InstanceActionsV221TestJSON-1322380366 tempest-InstanceActionsV221TestJSON-1322380366-project-member] Lock "7bc63fdd-4b50-486d-9681-ca0baa08f7b3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.712s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 743.920251] env[63531]: DEBUG oslo_vmware.api [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117637, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.517208} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.920511] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] e7fb00a3-0d2e-4f54-950d-337307112d7a/e7fb00a3-0d2e-4f54-950d-337307112d7a.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 743.920723] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 743.920970] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cab47661-3309-4b55-a0ea-e5ca746ee421 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.930277] env[63531]: DEBUG oslo_vmware.api [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Waiting for the task: (returnval){ [ 743.930277] env[63531]: value = "task-1117646" [ 743.930277] env[63531]: _type = "Task" [ 743.930277] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.938959] env[63531]: DEBUG oslo_vmware.api [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117646, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.969511] env[63531]: DEBUG oslo_vmware.api [None req-75b88ae3-9411-41cb-88e2-c8261dabea8d tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52c4ff64-1a30-606d-be31-ccc7af768f29, 'name': SearchDatastore_Task, 'duration_secs': 0.012016} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.969903] env[63531]: DEBUG oslo_concurrency.lockutils [None req-75b88ae3-9411-41cb-88e2-c8261dabea8d tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 743.970254] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-75b88ae3-9411-41cb-88e2-c8261dabea8d tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] c3920e60-b27f-496b-b616-afb23ab0bc26/9d550399-588c-4bd5-8436-eba9b523bd13-rescue.vmdk. {{(pid=63531) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 743.970576] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-df01141b-fde2-4e49-b4bf-67a50b2fa7bd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.979490] env[63531]: DEBUG oslo_vmware.api [None req-75b88ae3-9411-41cb-88e2-c8261dabea8d tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Waiting for the task: (returnval){ [ 743.979490] env[63531]: value = "task-1117647" [ 743.979490] env[63531]: _type = "Task" [ 743.979490] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.987286] env[63531]: DEBUG oslo_vmware.api [None req-75b88ae3-9411-41cb-88e2-c8261dabea8d tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Task: {'id': task-1117647, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.205381] env[63531]: DEBUG oslo_vmware.api [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1117645, 'name': PowerOnVM_Task} progress is 94%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.233609] env[63531]: DEBUG nova.network.neutron [req-a3ed3488-ac22-4945-8c1a-95333dc73f3a req-1005fc31-4525-454d-a616-36ec616d6e6d service nova] [instance: 0af99240-fa7a-4eed-a729-a5ae98c41cf9] Updated VIF entry in instance network info cache for port c2356641-c89b-4610-ade8-8cc4e392021b. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 744.233971] env[63531]: DEBUG nova.network.neutron [req-a3ed3488-ac22-4945-8c1a-95333dc73f3a req-1005fc31-4525-454d-a616-36ec616d6e6d service nova] [instance: 0af99240-fa7a-4eed-a729-a5ae98c41cf9] Updating instance_info_cache with network_info: [{"id": "c2356641-c89b-4610-ade8-8cc4e392021b", "address": "fa:16:3e:46:eb:a0", "network": {"id": "5439232a-ec6d-4184-a9f1-f7be5c1526bc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.186", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "54fbef8a6ee0488f820706d2036e9c99", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f267bcdd-0daa-4337-9709-5fc060c267d8", "external-id": "nsx-vlan-transportzone-308", "segmentation_id": 308, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2356641-c8", "ovs_interfaceid": "c2356641-c89b-4610-ade8-8cc4e392021b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 744.284303] env[63531]: DEBUG oslo_vmware.api [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52dda030-d3b8-bc1e-b4b8-0fd4dc38877d, 'name': SearchDatastore_Task, 'duration_secs': 0.009453} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.284561] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 744.284746] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] [instance: 0af99240-fa7a-4eed-a729-a5ae98c41cf9] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 744.284967] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 744.285133] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 744.285312] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 744.285581] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b2f84cef-81a4-461b-a1fb-1c3f95e267a3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.295307] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 744.295391] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 744.296193] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a949294b-caad-4bf2-9569-efd6c9b72849 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.304491] env[63531]: DEBUG oslo_vmware.api [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Waiting for the task: (returnval){ [ 744.304491] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]529fb2f0-7aab-6b0a-7d87-ffb1115ab1dd" [ 744.304491] env[63531]: _type = "Task" [ 744.304491] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.311110] env[63531]: DEBUG oslo_concurrency.lockutils [req-dcce3a33-952e-4d14-9782-4542fbc125a8 req-a729f9d1-7965-4630-864b-f8a3a204d939 service nova] Releasing lock "refresh_cache-410ddf7a-e45b-4769-a3db-5363ce2096a7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 744.314481] env[63531]: DEBUG oslo_vmware.api [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]529fb2f0-7aab-6b0a-7d87-ffb1115ab1dd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.441202] env[63531]: DEBUG oslo_vmware.api [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117646, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074145} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.441202] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 744.442125] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83849218-b33e-4bf4-82c0-77111d37b64f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.470638] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Reconfiguring VM instance instance-0000002b to attach disk [datastore2] e7fb00a3-0d2e-4f54-950d-337307112d7a/e7fb00a3-0d2e-4f54-950d-337307112d7a.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 744.473816] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9946d329-c436-4145-a371-553c8aeeb4cc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.499227] env[63531]: DEBUG oslo_vmware.api [None req-75b88ae3-9411-41cb-88e2-c8261dabea8d tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Task: {'id': task-1117647, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.50736} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.501581] env[63531]: INFO nova.virt.vmwareapi.ds_util [None req-75b88ae3-9411-41cb-88e2-c8261dabea8d tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] c3920e60-b27f-496b-b616-afb23ab0bc26/9d550399-588c-4bd5-8436-eba9b523bd13-rescue.vmdk. [ 744.501581] env[63531]: DEBUG oslo_vmware.api [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Waiting for the task: (returnval){ [ 744.501581] env[63531]: value = "task-1117648" [ 744.501581] env[63531]: _type = "Task" [ 744.501581] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.501806] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb0391f3-e116-4f36-affc-14d2157217a3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.532647] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-75b88ae3-9411-41cb-88e2-c8261dabea8d tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Reconfiguring VM instance instance-00000028 to attach disk [datastore1] c3920e60-b27f-496b-b616-afb23ab0bc26/9d550399-588c-4bd5-8436-eba9b523bd13-rescue.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 744.538782] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-368ca3ba-f8b9-4dbe-bc7e-687234ebde36 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.552597] env[63531]: DEBUG oslo_vmware.api [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117648, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.559220] env[63531]: DEBUG oslo_vmware.api [None req-75b88ae3-9411-41cb-88e2-c8261dabea8d tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Waiting for the task: (returnval){ [ 744.559220] env[63531]: value = "task-1117649" [ 744.559220] env[63531]: _type = "Task" [ 744.559220] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.567281] env[63531]: DEBUG oslo_vmware.api [None req-75b88ae3-9411-41cb-88e2-c8261dabea8d tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Task: {'id': task-1117649, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.705076] env[63531]: DEBUG oslo_vmware.api [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1117645, 'name': PowerOnVM_Task, 'duration_secs': 0.677101} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.705328] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 744.705524] env[63531]: INFO nova.compute.manager [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Took 8.96 seconds to spawn the instance on the hypervisor. [ 744.705702] env[63531]: DEBUG nova.compute.manager [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 744.706490] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88d5b980-c30e-4513-aa0f-1c5738fc630f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.739837] env[63531]: DEBUG oslo_concurrency.lockutils [req-a3ed3488-ac22-4945-8c1a-95333dc73f3a req-1005fc31-4525-454d-a616-36ec616d6e6d service nova] Releasing lock "refresh_cache-0af99240-fa7a-4eed-a729-a5ae98c41cf9" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 744.818251] env[63531]: DEBUG oslo_vmware.api [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]529fb2f0-7aab-6b0a-7d87-ffb1115ab1dd, 'name': SearchDatastore_Task, 'duration_secs': 0.017671} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.819067] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b5e59ba3-fabd-4dd5-9104-76bac5c6e326 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.824681] env[63531]: DEBUG oslo_vmware.api [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Waiting for the task: (returnval){ [ 744.824681] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52878ff7-a8d0-56b5-dd1b-ad2752e1e975" [ 744.824681] env[63531]: _type = "Task" [ 744.824681] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.836115] env[63531]: DEBUG oslo_vmware.api [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52878ff7-a8d0-56b5-dd1b-ad2752e1e975, 'name': SearchDatastore_Task} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.836390] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 744.836745] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] 0af99240-fa7a-4eed-a729-a5ae98c41cf9/0af99240-fa7a-4eed-a729-a5ae98c41cf9.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 744.836999] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-54245dbc-be58-434a-a2d6-60b53a0425ae {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.846772] env[63531]: DEBUG oslo_vmware.api [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Waiting for the task: (returnval){ [ 744.846772] env[63531]: value = "task-1117650" [ 744.846772] env[63531]: _type = "Task" [ 744.846772] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.855311] env[63531]: DEBUG oslo_vmware.api [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Task: {'id': task-1117650, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.861445] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0776b4a1-e013-455a-b465-942ab4bac480 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.868919] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8598ad34-3391-4798-9302-35b2b1b66df9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.906096] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7985dec7-5e75-41c8-88b9-d2613827216e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.915054] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57a1d927-0b3e-444e-8b9c-5f9e16c3b0fb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.929358] env[63531]: DEBUG nova.compute.provider_tree [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 745.015138] env[63531]: DEBUG oslo_vmware.api [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117648, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.070991] env[63531]: DEBUG oslo_vmware.api [None req-75b88ae3-9411-41cb-88e2-c8261dabea8d tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Task: {'id': task-1117649, 'name': ReconfigVM_Task, 'duration_secs': 0.4969} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.070991] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-75b88ae3-9411-41cb-88e2-c8261dabea8d tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Reconfigured VM instance instance-00000028 to attach disk [datastore1] c3920e60-b27f-496b-b616-afb23ab0bc26/9d550399-588c-4bd5-8436-eba9b523bd13-rescue.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 745.070991] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-764af272-4115-4fbf-a570-2b674dad174a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.105116] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1187ab2c-67d8-4a87-b852-fa714c073fbe {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.124425] env[63531]: DEBUG oslo_vmware.api [None req-75b88ae3-9411-41cb-88e2-c8261dabea8d tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Waiting for the task: (returnval){ [ 745.124425] env[63531]: value = "task-1117651" [ 745.124425] env[63531]: _type = "Task" [ 745.124425] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.137502] env[63531]: DEBUG oslo_vmware.api [None req-75b88ae3-9411-41cb-88e2-c8261dabea8d tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Task: {'id': task-1117651, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.228643] env[63531]: INFO nova.compute.manager [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Took 28.43 seconds to build instance. [ 745.359655] env[63531]: DEBUG oslo_vmware.api [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Task: {'id': task-1117650, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.509553} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.359905] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] 0af99240-fa7a-4eed-a729-a5ae98c41cf9/0af99240-fa7a-4eed-a729-a5ae98c41cf9.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 745.360129] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] [instance: 0af99240-fa7a-4eed-a729-a5ae98c41cf9] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 745.360433] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9bcd174f-2001-4198-8867-23bd51728211 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.368057] env[63531]: DEBUG oslo_vmware.api [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Waiting for the task: (returnval){ [ 745.368057] env[63531]: value = "task-1117652" [ 745.368057] env[63531]: _type = "Task" [ 745.368057] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.383012] env[63531]: DEBUG oslo_vmware.api [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Task: {'id': task-1117652, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.432301] env[63531]: DEBUG nova.scheduler.client.report [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 745.515175] env[63531]: DEBUG oslo_vmware.api [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117648, 'name': ReconfigVM_Task, 'duration_secs': 0.677722} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.515175] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Reconfigured VM instance instance-0000002b to attach disk [datastore2] e7fb00a3-0d2e-4f54-950d-337307112d7a/e7fb00a3-0d2e-4f54-950d-337307112d7a.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 745.516105] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4b71f83c-425b-4eb1-bfa0-a77b3e271126 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.522602] env[63531]: DEBUG oslo_vmware.api [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Waiting for the task: (returnval){ [ 745.522602] env[63531]: value = "task-1117653" [ 745.522602] env[63531]: _type = "Task" [ 745.522602] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.532992] env[63531]: DEBUG oslo_vmware.api [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117653, 'name': Rename_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.635756] env[63531]: DEBUG oslo_vmware.api [None req-75b88ae3-9411-41cb-88e2-c8261dabea8d tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Task: {'id': task-1117651, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.733542] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5f63bbbe-ea5a-412b-a8d5-e8309416c3aa tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Lock "1ca7b851-2a06-4181-8271-58aafcd322d6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 128.360s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 745.879480] env[63531]: DEBUG oslo_vmware.api [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Task: {'id': task-1117652, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.253823} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.879817] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] [instance: 0af99240-fa7a-4eed-a729-a5ae98c41cf9] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 745.880693] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8784fc9-f634-407c-b357-0ecaac077133 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.912399] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] [instance: 0af99240-fa7a-4eed-a729-a5ae98c41cf9] Reconfiguring VM instance instance-0000002c to attach disk [datastore2] 0af99240-fa7a-4eed-a729-a5ae98c41cf9/0af99240-fa7a-4eed-a729-a5ae98c41cf9.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 745.912399] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1b769f91-c1fa-4249-b0e4-46cf50d19e6c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.932519] env[63531]: DEBUG oslo_vmware.api [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Waiting for the task: (returnval){ [ 745.932519] env[63531]: value = "task-1117654" [ 745.932519] env[63531]: _type = "Task" [ 745.932519] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.943238] env[63531]: DEBUG oslo_concurrency.lockutils [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.585s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 745.943238] env[63531]: DEBUG nova.compute.manager [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 745.944811] env[63531]: DEBUG oslo_vmware.api [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Task: {'id': task-1117654, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.945326] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.557s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 745.947354] env[63531]: INFO nova.compute.claims [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 746.033406] env[63531]: DEBUG oslo_vmware.api [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117653, 'name': Rename_Task, 'duration_secs': 0.241976} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.033678] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 746.033922] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9496360e-dc41-4e2f-a479-c170b2028063 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.040892] env[63531]: DEBUG oslo_vmware.api [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Waiting for the task: (returnval){ [ 746.040892] env[63531]: value = "task-1117655" [ 746.040892] env[63531]: _type = "Task" [ 746.040892] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.049801] env[63531]: DEBUG oslo_vmware.api [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117655, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.136553] env[63531]: DEBUG oslo_vmware.api [None req-75b88ae3-9411-41cb-88e2-c8261dabea8d tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Task: {'id': task-1117651, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.237495] env[63531]: DEBUG nova.compute.manager [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 746.449158] env[63531]: DEBUG nova.compute.utils [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 746.450501] env[63531]: DEBUG oslo_vmware.api [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Task: {'id': task-1117654, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.450977] env[63531]: DEBUG nova.compute.manager [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 746.451448] env[63531]: DEBUG nova.network.neutron [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 746.556939] env[63531]: DEBUG oslo_vmware.api [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117655, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.558714] env[63531]: DEBUG nova.policy [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '07142a4db1924c1484eb60ee55aec06c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cf90f907de64448ca9c01029f694ed6c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 746.646803] env[63531]: DEBUG oslo_vmware.api [None req-75b88ae3-9411-41cb-88e2-c8261dabea8d tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Task: {'id': task-1117651, 'name': ReconfigVM_Task, 'duration_secs': 1.252129} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.647148] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-75b88ae3-9411-41cb-88e2-c8261dabea8d tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 746.647427] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cb78eab2-4012-4702-852c-8c84040ebb72 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.655296] env[63531]: DEBUG oslo_vmware.api [None req-75b88ae3-9411-41cb-88e2-c8261dabea8d tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Waiting for the task: (returnval){ [ 746.655296] env[63531]: value = "task-1117656" [ 746.655296] env[63531]: _type = "Task" [ 746.655296] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.664825] env[63531]: DEBUG oslo_vmware.api [None req-75b88ae3-9411-41cb-88e2-c8261dabea8d tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Task: {'id': task-1117656, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.758275] env[63531]: DEBUG oslo_concurrency.lockutils [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 746.947149] env[63531]: DEBUG oslo_vmware.api [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Task: {'id': task-1117654, 'name': ReconfigVM_Task, 'duration_secs': 0.955149} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.950821] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] [instance: 0af99240-fa7a-4eed-a729-a5ae98c41cf9] Reconfigured VM instance instance-0000002c to attach disk [datastore2] 0af99240-fa7a-4eed-a729-a5ae98c41cf9/0af99240-fa7a-4eed-a729-a5ae98c41cf9.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 746.951953] env[63531]: DEBUG nova.compute.manager [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 746.957407] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dc3a45a1-5309-41c8-a21c-3af2c875b0f0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.967487] env[63531]: DEBUG oslo_vmware.api [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Waiting for the task: (returnval){ [ 746.967487] env[63531]: value = "task-1117657" [ 746.967487] env[63531]: _type = "Task" [ 746.967487] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.978319] env[63531]: DEBUG oslo_vmware.api [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Task: {'id': task-1117657, 'name': Rename_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.011414] env[63531]: DEBUG nova.network.neutron [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74] Successfully created port: 8a0490e2-e2d3-4316-9682-7d9a24122671 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 747.058826] env[63531]: DEBUG oslo_vmware.api [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117655, 'name': PowerOnVM_Task, 'duration_secs': 0.891499} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.059110] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 747.059867] env[63531]: INFO nova.compute.manager [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Took 8.76 seconds to spawn the instance on the hypervisor. [ 747.059867] env[63531]: DEBUG nova.compute.manager [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 747.060462] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bebfc386-aed8-4348-bc6c-087e7fc746b9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.171917] env[63531]: DEBUG oslo_vmware.api [None req-75b88ae3-9411-41cb-88e2-c8261dabea8d tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Task: {'id': task-1117656, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.179677] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Acquiring lock "531e1852-0c67-4b4d-a0ca-749ac438e79a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 747.180485] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Lock "531e1852-0c67-4b4d-a0ca-749ac438e79a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 747.423194] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b91f8a8c-9812-46dc-b3a2-17bed490ae67 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.431111] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af80933f-812a-4fd4-a7b6-707d541b11c1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.470740] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e6a0980-c68c-4377-996f-c94cc898dae5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.487026] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcffb0bf-438e-43f0-b432-3419e9df55f4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.491585] env[63531]: DEBUG oslo_vmware.api [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Task: {'id': task-1117657, 'name': Rename_Task, 'duration_secs': 0.204783} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.491950] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] [instance: 0af99240-fa7a-4eed-a729-a5ae98c41cf9] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 747.492708] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-39d8d651-cfc1-470c-a84d-02b882251797 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.506566] env[63531]: DEBUG nova.compute.provider_tree [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 747.514482] env[63531]: DEBUG oslo_vmware.api [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Waiting for the task: (returnval){ [ 747.514482] env[63531]: value = "task-1117658" [ 747.514482] env[63531]: _type = "Task" [ 747.514482] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.526123] env[63531]: DEBUG oslo_vmware.api [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Task: {'id': task-1117658, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.582335] env[63531]: INFO nova.compute.manager [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Took 24.18 seconds to build instance. [ 747.668107] env[63531]: DEBUG oslo_vmware.api [None req-75b88ae3-9411-41cb-88e2-c8261dabea8d tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Task: {'id': task-1117656, 'name': PowerOnVM_Task, 'duration_secs': 0.548516} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.668385] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-75b88ae3-9411-41cb-88e2-c8261dabea8d tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 747.671187] env[63531]: DEBUG nova.compute.manager [None req-75b88ae3-9411-41cb-88e2-c8261dabea8d tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 747.672243] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb65dfe4-8c78-437a-bb67-13200f6381bc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.975335] env[63531]: DEBUG nova.compute.manager [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 748.006028] env[63531]: DEBUG nova.virt.hardware [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 748.006028] env[63531]: DEBUG nova.virt.hardware [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 748.006028] env[63531]: DEBUG nova.virt.hardware [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 748.006331] env[63531]: DEBUG nova.virt.hardware [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 748.006331] env[63531]: DEBUG nova.virt.hardware [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 748.006331] env[63531]: DEBUG nova.virt.hardware [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 748.006331] env[63531]: DEBUG nova.virt.hardware [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 748.006597] env[63531]: DEBUG nova.virt.hardware [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 748.006935] env[63531]: DEBUG nova.virt.hardware [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 748.007453] env[63531]: DEBUG nova.virt.hardware [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 748.007814] env[63531]: DEBUG nova.virt.hardware [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 748.009748] env[63531]: DEBUG nova.scheduler.client.report [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 748.014681] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-535bea23-314e-4434-b378-348e40823882 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.030445] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5de96c66-090d-4dc7-b8c5-f4d3fde04a9a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.036373] env[63531]: DEBUG oslo_vmware.api [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Task: {'id': task-1117658, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.084684] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ac7af755-5263-471d-b6f6-c94116d58c51 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Lock "e7fb00a3-0d2e-4f54-950d-337307112d7a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 126.440s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 748.249477] env[63531]: INFO nova.compute.manager [None req-9b7b17b4-8f5f-4f77-83b4-2a61816669ba tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Rescuing [ 748.249844] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9b7b17b4-8f5f-4f77-83b4-2a61816669ba tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Acquiring lock "refresh_cache-e7fb00a3-0d2e-4f54-950d-337307112d7a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 748.250066] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9b7b17b4-8f5f-4f77-83b4-2a61816669ba tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Acquired lock "refresh_cache-e7fb00a3-0d2e-4f54-950d-337307112d7a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 748.250272] env[63531]: DEBUG nova.network.neutron [None req-9b7b17b4-8f5f-4f77-83b4-2a61816669ba tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 748.522026] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.574s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 748.522026] env[63531]: DEBUG nova.compute.manager [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 748.523297] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1a515fd1-1aca-4043-ae4d-827540148b79 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.726s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 748.523644] env[63531]: DEBUG nova.objects.instance [None req-1a515fd1-1aca-4043-ae4d-827540148b79 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Lazy-loading 'resources' on Instance uuid 0c948655-362d-4673-b3e6-3f28ec69ea3c {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 748.539801] env[63531]: DEBUG oslo_vmware.api [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Task: {'id': task-1117658, 'name': PowerOnVM_Task} progress is 79%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.594907] env[63531]: DEBUG nova.compute.manager [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 748.598948] env[63531]: DEBUG nova.compute.manager [req-47ed717f-2793-4634-a757-35be51050330 req-f27cf526-f4a9-46b2-9329-530b4fa93824 service nova] [instance: e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74] Received event network-vif-plugged-8a0490e2-e2d3-4316-9682-7d9a24122671 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 748.599397] env[63531]: DEBUG oslo_concurrency.lockutils [req-47ed717f-2793-4634-a757-35be51050330 req-f27cf526-f4a9-46b2-9329-530b4fa93824 service nova] Acquiring lock "e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 748.599729] env[63531]: DEBUG oslo_concurrency.lockutils [req-47ed717f-2793-4634-a757-35be51050330 req-f27cf526-f4a9-46b2-9329-530b4fa93824 service nova] Lock "e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 748.600025] env[63531]: DEBUG oslo_concurrency.lockutils [req-47ed717f-2793-4634-a757-35be51050330 req-f27cf526-f4a9-46b2-9329-530b4fa93824 service nova] Lock "e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 748.600298] env[63531]: DEBUG nova.compute.manager [req-47ed717f-2793-4634-a757-35be51050330 req-f27cf526-f4a9-46b2-9329-530b4fa93824 service nova] [instance: e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74] No waiting events found dispatching network-vif-plugged-8a0490e2-e2d3-4316-9682-7d9a24122671 {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 748.600568] env[63531]: WARNING nova.compute.manager [req-47ed717f-2793-4634-a757-35be51050330 req-f27cf526-f4a9-46b2-9329-530b4fa93824 service nova] [instance: e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74] Received unexpected event network-vif-plugged-8a0490e2-e2d3-4316-9682-7d9a24122671 for instance with vm_state building and task_state spawning. [ 748.785281] env[63531]: DEBUG nova.network.neutron [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74] Successfully updated port: 8a0490e2-e2d3-4316-9682-7d9a24122671 {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 749.024763] env[63531]: DEBUG nova.compute.utils [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 749.036528] env[63531]: DEBUG nova.compute.manager [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 749.036707] env[63531]: DEBUG nova.network.neutron [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 749.053643] env[63531]: DEBUG oslo_vmware.api [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Task: {'id': task-1117658, 'name': PowerOnVM_Task, 'duration_secs': 1.45357} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.054941] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] [instance: 0af99240-fa7a-4eed-a729-a5ae98c41cf9] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 749.055165] env[63531]: INFO nova.compute.manager [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] [instance: 0af99240-fa7a-4eed-a729-a5ae98c41cf9] Took 8.27 seconds to spawn the instance on the hypervisor. [ 749.055362] env[63531]: DEBUG nova.compute.manager [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] [instance: 0af99240-fa7a-4eed-a729-a5ae98c41cf9] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 749.056953] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e4df6e2-3461-400d-af83-abff5782fc88 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.122177] env[63531]: DEBUG oslo_concurrency.lockutils [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 749.292811] env[63531]: DEBUG oslo_concurrency.lockutils [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Acquiring lock "refresh_cache-e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 749.292914] env[63531]: DEBUG oslo_concurrency.lockutils [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Acquired lock "refresh_cache-e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.293044] env[63531]: DEBUG nova.network.neutron [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 749.298489] env[63531]: DEBUG nova.policy [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1af2cb839598417b897695b83e867058', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8bebde0e281d485ebdd70a1a3b613632', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 749.315459] env[63531]: DEBUG nova.network.neutron [None req-9b7b17b4-8f5f-4f77-83b4-2a61816669ba tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Updating instance_info_cache with network_info: [{"id": "6102475b-8e43-4b77-85c3-5b574383c06d", "address": "fa:16:3e:9c:74:35", "network": {"id": "dc416058-1c04-47ac-8b73-1557f7c33d1c", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1350689703-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "824ed405cc6e4868a70e82489e05e8ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "05b1253d-2b87-4158-9ff1-dafcf829f11f", "external-id": "nsx-vlan-transportzone-55", "segmentation_id": 55, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6102475b-8e", "ovs_interfaceid": "6102475b-8e43-4b77-85c3-5b574383c06d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.461849] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58a1b891-9378-421e-81da-c94947cea381 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.472198] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee0b0f98-728e-493e-acf3-f822fb199034 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.510422] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bccd1b7b-4532-4aea-9fb5-79c81ea1d605 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.518623] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee74d833-1bda-4611-85eb-d40a954432bc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.534550] env[63531]: DEBUG nova.compute.provider_tree [None req-1a515fd1-1aca-4043-ae4d-827540148b79 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 749.537431] env[63531]: DEBUG nova.compute.manager [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 749.585366] env[63531]: INFO nova.compute.manager [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] [instance: 0af99240-fa7a-4eed-a729-a5ae98c41cf9] Took 26.01 seconds to build instance. [ 749.772819] env[63531]: DEBUG nova.network.neutron [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Successfully created port: 4c96ebbd-98fa-4d12-a388-b9d270b7aaae {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 749.816611] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9b7b17b4-8f5f-4f77-83b4-2a61816669ba tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Releasing lock "refresh_cache-e7fb00a3-0d2e-4f54-950d-337307112d7a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 749.855352] env[63531]: DEBUG nova.network.neutron [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 750.038902] env[63531]: DEBUG nova.scheduler.client.report [None req-1a515fd1-1aca-4043-ae4d-827540148b79 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 750.088065] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3aa9759d-816a-426a-9483-01f0df54ca68 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Lock "0af99240-fa7a-4eed-a729-a5ae98c41cf9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 128.385s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 750.231284] env[63531]: DEBUG nova.network.neutron [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Successfully created port: 5765492b-2d11-438b-b9a4-7514ce2ac73e {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 750.358915] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b7b17b4-8f5f-4f77-83b4-2a61816669ba tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 750.358915] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c125300b-169f-463d-82d4-7c202e60cbd9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.368408] env[63531]: DEBUG oslo_vmware.api [None req-9b7b17b4-8f5f-4f77-83b4-2a61816669ba tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Waiting for the task: (returnval){ [ 750.368408] env[63531]: value = "task-1117659" [ 750.368408] env[63531]: _type = "Task" [ 750.368408] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.383335] env[63531]: DEBUG oslo_vmware.api [None req-9b7b17b4-8f5f-4f77-83b4-2a61816669ba tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117659, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.456047] env[63531]: DEBUG nova.network.neutron [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74] Updating instance_info_cache with network_info: [{"id": "8a0490e2-e2d3-4316-9682-7d9a24122671", "address": "fa:16:3e:bc:66:de", "network": {"id": "6d1097c7-db4f-4446-a36f-fd02a69df0e2", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-917061070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cf90f907de64448ca9c01029f694ed6c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a88b707-352e-4be7-b1d6-ad6074b40ed9", "external-id": "nsx-vlan-transportzone-789", "segmentation_id": 789, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8a0490e2-e2", "ovs_interfaceid": "8a0490e2-e2d3-4316-9682-7d9a24122671", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.544204] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1a515fd1-1aca-4043-ae4d-827540148b79 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.021s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 750.549409] env[63531]: DEBUG oslo_concurrency.lockutils [None req-524a4683-a8e1-4d4b-87d9-f141433baa74 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.222s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 750.549699] env[63531]: DEBUG nova.objects.instance [None req-524a4683-a8e1-4d4b-87d9-f141433baa74 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Lazy-loading 'resources' on Instance uuid f0276e34-1de3-482a-82c7-1439d40fd85e {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 750.552046] env[63531]: DEBUG nova.compute.manager [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 750.585512] env[63531]: DEBUG nova.virt.hardware [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 750.585798] env[63531]: DEBUG nova.virt.hardware [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 750.586022] env[63531]: DEBUG nova.virt.hardware [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 750.586270] env[63531]: DEBUG nova.virt.hardware [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 750.586432] env[63531]: DEBUG nova.virt.hardware [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 750.586567] env[63531]: DEBUG nova.virt.hardware [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 750.586778] env[63531]: DEBUG nova.virt.hardware [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 750.587772] env[63531]: DEBUG nova.virt.hardware [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 750.587772] env[63531]: DEBUG nova.virt.hardware [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 750.587772] env[63531]: DEBUG nova.virt.hardware [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 750.587772] env[63531]: DEBUG nova.virt.hardware [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 750.588795] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c897cc31-a1cc-4544-a221-9d6e5ad3c345 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.594154] env[63531]: INFO nova.scheduler.client.report [None req-1a515fd1-1aca-4043-ae4d-827540148b79 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Deleted allocations for instance 0c948655-362d-4673-b3e6-3f28ec69ea3c [ 750.595140] env[63531]: DEBUG nova.compute.manager [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 750.605533] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44c44d4f-5e5c-4ca0-a9fc-a8b6b72e5547 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.812220] env[63531]: DEBUG nova.network.neutron [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Successfully created port: 4da738f4-b2e9-40c4-9a02-43b6ed658cfa {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 750.879312] env[63531]: DEBUG oslo_vmware.api [None req-9b7b17b4-8f5f-4f77-83b4-2a61816669ba tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117659, 'name': PowerOffVM_Task, 'duration_secs': 0.231718} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.879582] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b7b17b4-8f5f-4f77-83b4-2a61816669ba tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 750.884023] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4112041f-2a46-411c-91b5-d9da99984bb3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.899255] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4233cade-d65b-49fc-a10b-840a1b4e154e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.937509] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b7b17b4-8f5f-4f77-83b4-2a61816669ba tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 750.938093] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-50842bb9-0320-44c6-bf1d-4f3792f48b3d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.948371] env[63531]: DEBUG oslo_vmware.api [None req-9b7b17b4-8f5f-4f77-83b4-2a61816669ba tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Waiting for the task: (returnval){ [ 750.948371] env[63531]: value = "task-1117660" [ 750.948371] env[63531]: _type = "Task" [ 750.948371] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.955128] env[63531]: DEBUG oslo_vmware.api [None req-9b7b17b4-8f5f-4f77-83b4-2a61816669ba tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117660, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.962019] env[63531]: DEBUG oslo_concurrency.lockutils [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Releasing lock "refresh_cache-e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 750.962019] env[63531]: DEBUG nova.compute.manager [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74] Instance network_info: |[{"id": "8a0490e2-e2d3-4316-9682-7d9a24122671", "address": "fa:16:3e:bc:66:de", "network": {"id": "6d1097c7-db4f-4446-a36f-fd02a69df0e2", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-917061070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cf90f907de64448ca9c01029f694ed6c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a88b707-352e-4be7-b1d6-ad6074b40ed9", "external-id": "nsx-vlan-transportzone-789", "segmentation_id": 789, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8a0490e2-e2", "ovs_interfaceid": "8a0490e2-e2d3-4316-9682-7d9a24122671", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 750.962313] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bc:66:de', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0a88b707-352e-4be7-b1d6-ad6074b40ed9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8a0490e2-e2d3-4316-9682-7d9a24122671', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 750.968549] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Creating folder: Project (cf90f907de64448ca9c01029f694ed6c). Parent ref: group-v244585. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 750.968947] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d78f782e-7e00-4b6b-9363-15ac662c55dc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.981609] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Created folder: Project (cf90f907de64448ca9c01029f694ed6c) in parent group-v244585. [ 750.981814] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Creating folder: Instances. Parent ref: group-v244637. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 750.982067] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-652df12b-775a-4bf7-ab91-8d4aca25d5f0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.993154] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Created folder: Instances in parent group-v244637. [ 750.993154] env[63531]: DEBUG oslo.service.loopingcall [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 750.993154] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 750.993154] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-34c1fe0b-bb0a-46e0-b2b5-336a99de871c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.008046] env[63531]: DEBUG nova.compute.manager [req-02089103-b68f-4ab6-add6-75252eee6fba req-6a5eeab8-88af-4e52-8fb4-5d82381b6374 service nova] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Received event network-changed-91733d5a-84d6-48df-ba4d-7e362a642282 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 751.008261] env[63531]: DEBUG nova.compute.manager [req-02089103-b68f-4ab6-add6-75252eee6fba req-6a5eeab8-88af-4e52-8fb4-5d82381b6374 service nova] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Refreshing instance network info cache due to event network-changed-91733d5a-84d6-48df-ba4d-7e362a642282. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 751.008474] env[63531]: DEBUG oslo_concurrency.lockutils [req-02089103-b68f-4ab6-add6-75252eee6fba req-6a5eeab8-88af-4e52-8fb4-5d82381b6374 service nova] Acquiring lock "refresh_cache-c3920e60-b27f-496b-b616-afb23ab0bc26" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 751.008614] env[63531]: DEBUG oslo_concurrency.lockutils [req-02089103-b68f-4ab6-add6-75252eee6fba req-6a5eeab8-88af-4e52-8fb4-5d82381b6374 service nova] Acquired lock "refresh_cache-c3920e60-b27f-496b-b616-afb23ab0bc26" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 751.008780] env[63531]: DEBUG nova.network.neutron [req-02089103-b68f-4ab6-add6-75252eee6fba req-6a5eeab8-88af-4e52-8fb4-5d82381b6374 service nova] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Refreshing network info cache for port 91733d5a-84d6-48df-ba4d-7e362a642282 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 751.015037] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 751.015037] env[63531]: value = "task-1117663" [ 751.015037] env[63531]: _type = "Task" [ 751.015037] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.023027] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117663, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.135342] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 751.136291] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1a515fd1-1aca-4043-ae4d-827540148b79 tempest-ServerShowV257Test-1260117185 tempest-ServerShowV257Test-1260117185-project-member] Lock "0c948655-362d-4673-b3e6-3f28ec69ea3c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.058s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 751.274733] env[63531]: DEBUG nova.compute.manager [req-72f3e474-a36d-4868-92f9-edc6560f0576 req-6c41cd46-3c8d-4d8f-8d14-90c8235e08d1 service nova] [instance: e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74] Received event network-changed-8a0490e2-e2d3-4316-9682-7d9a24122671 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 751.275994] env[63531]: DEBUG nova.compute.manager [req-72f3e474-a36d-4868-92f9-edc6560f0576 req-6c41cd46-3c8d-4d8f-8d14-90c8235e08d1 service nova] [instance: e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74] Refreshing instance network info cache due to event network-changed-8a0490e2-e2d3-4316-9682-7d9a24122671. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 751.275994] env[63531]: DEBUG oslo_concurrency.lockutils [req-72f3e474-a36d-4868-92f9-edc6560f0576 req-6c41cd46-3c8d-4d8f-8d14-90c8235e08d1 service nova] Acquiring lock "refresh_cache-e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 751.275994] env[63531]: DEBUG oslo_concurrency.lockutils [req-72f3e474-a36d-4868-92f9-edc6560f0576 req-6c41cd46-3c8d-4d8f-8d14-90c8235e08d1 service nova] Acquired lock "refresh_cache-e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 751.275994] env[63531]: DEBUG nova.network.neutron [req-72f3e474-a36d-4868-92f9-edc6560f0576 req-6c41cd46-3c8d-4d8f-8d14-90c8235e08d1 service nova] [instance: e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74] Refreshing network info cache for port 8a0490e2-e2d3-4316-9682-7d9a24122671 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 751.460830] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b7b17b4-8f5f-4f77-83b4-2a61816669ba tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] VM already powered off {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 751.461689] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-9b7b17b4-8f5f-4f77-83b4-2a61816669ba tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 751.461689] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9b7b17b4-8f5f-4f77-83b4-2a61816669ba tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 751.461689] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9b7b17b4-8f5f-4f77-83b4-2a61816669ba tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 751.461826] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b7b17b4-8f5f-4f77-83b4-2a61816669ba tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 751.462595] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0644f806-b170-41e9-9a3d-e1471e4b0137 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.486344] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b7b17b4-8f5f-4f77-83b4-2a61816669ba tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 751.486344] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-9b7b17b4-8f5f-4f77-83b4-2a61816669ba tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 751.486744] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c6a32515-e48f-460b-bb16-8e0d82733689 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.499607] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12efc090-07d5-4c11-b842-8cf422b34ba2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.503934] env[63531]: DEBUG oslo_vmware.api [None req-9b7b17b4-8f5f-4f77-83b4-2a61816669ba tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Waiting for the task: (returnval){ [ 751.503934] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]524f5fec-0923-70ac-a1ec-158d735fefbc" [ 751.503934] env[63531]: _type = "Task" [ 751.503934] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.513994] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a918596-b252-4681-87c6-d0cb1e12e04d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.522269] env[63531]: DEBUG oslo_vmware.api [None req-9b7b17b4-8f5f-4f77-83b4-2a61816669ba tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]524f5fec-0923-70ac-a1ec-158d735fefbc, 'name': SearchDatastore_Task, 'duration_secs': 0.01086} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.528068] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a64739b6-1e66-4e3e-8a27-f9a8e0fb74e9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.558403] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b7e6593-27d5-40fa-93f7-5b1f2d33dea6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.565264] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117663, 'name': CreateVM_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.565971] env[63531]: DEBUG oslo_vmware.api [None req-9b7b17b4-8f5f-4f77-83b4-2a61816669ba tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Waiting for the task: (returnval){ [ 751.565971] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52ea9f4a-e78c-7244-80e1-c69dec8ef231" [ 751.565971] env[63531]: _type = "Task" [ 751.565971] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.572671] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e256a1b2-388d-4a1d-bf75-3efbae782678 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.582562] env[63531]: DEBUG oslo_vmware.api [None req-9b7b17b4-8f5f-4f77-83b4-2a61816669ba tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52ea9f4a-e78c-7244-80e1-c69dec8ef231, 'name': SearchDatastore_Task, 'duration_secs': 0.009274} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.583231] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9b7b17b4-8f5f-4f77-83b4-2a61816669ba tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 751.583231] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b7b17b4-8f5f-4f77-83b4-2a61816669ba tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] e7fb00a3-0d2e-4f54-950d-337307112d7a/9d550399-588c-4bd5-8436-eba9b523bd13-rescue.vmdk. {{(pid=63531) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 751.583757] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-67732572-6077-4c1d-ac21-fc0557d942f6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.595014] env[63531]: DEBUG nova.compute.provider_tree [None req-524a4683-a8e1-4d4b-87d9-f141433baa74 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 751.604775] env[63531]: DEBUG oslo_vmware.api [None req-9b7b17b4-8f5f-4f77-83b4-2a61816669ba tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Waiting for the task: (returnval){ [ 751.604775] env[63531]: value = "task-1117664" [ 751.604775] env[63531]: _type = "Task" [ 751.604775] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.617486] env[63531]: DEBUG oslo_vmware.api [None req-9b7b17b4-8f5f-4f77-83b4-2a61816669ba tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117664, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.000726] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a56197e5-b9a2-4ba5-a23f-e676b6cc2b13 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Acquiring lock "0af99240-fa7a-4eed-a729-a5ae98c41cf9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 752.000726] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a56197e5-b9a2-4ba5-a23f-e676b6cc2b13 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Lock "0af99240-fa7a-4eed-a729-a5ae98c41cf9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 752.001069] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a56197e5-b9a2-4ba5-a23f-e676b6cc2b13 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Acquiring lock "0af99240-fa7a-4eed-a729-a5ae98c41cf9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 752.001069] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a56197e5-b9a2-4ba5-a23f-e676b6cc2b13 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Lock "0af99240-fa7a-4eed-a729-a5ae98c41cf9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 752.001200] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a56197e5-b9a2-4ba5-a23f-e676b6cc2b13 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Lock "0af99240-fa7a-4eed-a729-a5ae98c41cf9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 752.003861] env[63531]: INFO nova.compute.manager [None req-a56197e5-b9a2-4ba5-a23f-e676b6cc2b13 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] [instance: 0af99240-fa7a-4eed-a729-a5ae98c41cf9] Terminating instance [ 752.006222] env[63531]: DEBUG nova.compute.manager [None req-a56197e5-b9a2-4ba5-a23f-e676b6cc2b13 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] [instance: 0af99240-fa7a-4eed-a729-a5ae98c41cf9] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 752.006616] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a56197e5-b9a2-4ba5-a23f-e676b6cc2b13 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] [instance: 0af99240-fa7a-4eed-a729-a5ae98c41cf9] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 752.007604] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bb89ec9-7859-4eff-bb42-adebe622fb05 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.016819] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-a56197e5-b9a2-4ba5-a23f-e676b6cc2b13 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] [instance: 0af99240-fa7a-4eed-a729-a5ae98c41cf9] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 752.017167] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e7d5a01c-25db-4ab9-859b-d16803090e37 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.029964] env[63531]: DEBUG oslo_vmware.api [None req-a56197e5-b9a2-4ba5-a23f-e676b6cc2b13 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Waiting for the task: (returnval){ [ 752.029964] env[63531]: value = "task-1117665" [ 752.029964] env[63531]: _type = "Task" [ 752.029964] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.039329] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117663, 'name': CreateVM_Task, 'duration_secs': 0.928373} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.040109] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 752.041032] env[63531]: DEBUG oslo_concurrency.lockutils [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 752.041432] env[63531]: DEBUG oslo_concurrency.lockutils [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.041870] env[63531]: DEBUG oslo_concurrency.lockutils [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 752.045826] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-83907fd9-b8e4-4137-ab60-00afc01be793 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.049294] env[63531]: DEBUG oslo_vmware.api [None req-a56197e5-b9a2-4ba5-a23f-e676b6cc2b13 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Task: {'id': task-1117665, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.055226] env[63531]: DEBUG oslo_vmware.api [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Waiting for the task: (returnval){ [ 752.055226] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]520631b0-7dd3-f4b5-534f-587f274c4eee" [ 752.055226] env[63531]: _type = "Task" [ 752.055226] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.065503] env[63531]: DEBUG oslo_vmware.api [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]520631b0-7dd3-f4b5-534f-587f274c4eee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.103020] env[63531]: DEBUG nova.scheduler.client.report [None req-524a4683-a8e1-4d4b-87d9-f141433baa74 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 752.116199] env[63531]: DEBUG oslo_vmware.api [None req-9b7b17b4-8f5f-4f77-83b4-2a61816669ba tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117664, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.493551} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.116631] env[63531]: INFO nova.virt.vmwareapi.ds_util [None req-9b7b17b4-8f5f-4f77-83b4-2a61816669ba tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] e7fb00a3-0d2e-4f54-950d-337307112d7a/9d550399-588c-4bd5-8436-eba9b523bd13-rescue.vmdk. [ 752.117585] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ac9ea80-5937-47b4-aad8-f1ec282170f2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.149035] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-9b7b17b4-8f5f-4f77-83b4-2a61816669ba tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Reconfiguring VM instance instance-0000002b to attach disk [datastore2] e7fb00a3-0d2e-4f54-950d-337307112d7a/9d550399-588c-4bd5-8436-eba9b523bd13-rescue.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 752.155145] env[63531]: DEBUG nova.network.neutron [req-02089103-b68f-4ab6-add6-75252eee6fba req-6a5eeab8-88af-4e52-8fb4-5d82381b6374 service nova] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Updated VIF entry in instance network info cache for port 91733d5a-84d6-48df-ba4d-7e362a642282. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 752.155890] env[63531]: DEBUG nova.network.neutron [req-02089103-b68f-4ab6-add6-75252eee6fba req-6a5eeab8-88af-4e52-8fb4-5d82381b6374 service nova] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Updating instance_info_cache with network_info: [{"id": "91733d5a-84d6-48df-ba4d-7e362a642282", "address": "fa:16:3e:df:a1:31", "network": {"id": "d28b2087-329c-4917-aa6e-83ab83a63a74", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1054369059-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a634a61020b94d62a8fb37a3f96acc10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91733d5a-84", "ovs_interfaceid": "91733d5a-84d6-48df-ba4d-7e362a642282", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 752.159463] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-18ceca02-87ab-4e20-984f-ac208da0be91 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.174043] env[63531]: DEBUG oslo_concurrency.lockutils [req-02089103-b68f-4ab6-add6-75252eee6fba req-6a5eeab8-88af-4e52-8fb4-5d82381b6374 service nova] Releasing lock "refresh_cache-c3920e60-b27f-496b-b616-afb23ab0bc26" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 752.181563] env[63531]: DEBUG oslo_vmware.api [None req-9b7b17b4-8f5f-4f77-83b4-2a61816669ba tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Waiting for the task: (returnval){ [ 752.181563] env[63531]: value = "task-1117666" [ 752.181563] env[63531]: _type = "Task" [ 752.181563] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.192579] env[63531]: DEBUG oslo_vmware.api [None req-9b7b17b4-8f5f-4f77-83b4-2a61816669ba tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117666, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.373284] env[63531]: DEBUG nova.network.neutron [req-72f3e474-a36d-4868-92f9-edc6560f0576 req-6c41cd46-3c8d-4d8f-8d14-90c8235e08d1 service nova] [instance: e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74] Updated VIF entry in instance network info cache for port 8a0490e2-e2d3-4316-9682-7d9a24122671. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 752.373673] env[63531]: DEBUG nova.network.neutron [req-72f3e474-a36d-4868-92f9-edc6560f0576 req-6c41cd46-3c8d-4d8f-8d14-90c8235e08d1 service nova] [instance: e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74] Updating instance_info_cache with network_info: [{"id": "8a0490e2-e2d3-4316-9682-7d9a24122671", "address": "fa:16:3e:bc:66:de", "network": {"id": "6d1097c7-db4f-4446-a36f-fd02a69df0e2", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-917061070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cf90f907de64448ca9c01029f694ed6c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a88b707-352e-4be7-b1d6-ad6074b40ed9", "external-id": "nsx-vlan-transportzone-789", "segmentation_id": 789, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8a0490e2-e2", "ovs_interfaceid": "8a0490e2-e2d3-4316-9682-7d9a24122671", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 752.542568] env[63531]: DEBUG oslo_vmware.api [None req-a56197e5-b9a2-4ba5-a23f-e676b6cc2b13 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Task: {'id': task-1117665, 'name': PowerOffVM_Task, 'duration_secs': 0.261531} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.542568] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-a56197e5-b9a2-4ba5-a23f-e676b6cc2b13 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] [instance: 0af99240-fa7a-4eed-a729-a5ae98c41cf9] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 752.542568] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a56197e5-b9a2-4ba5-a23f-e676b6cc2b13 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] [instance: 0af99240-fa7a-4eed-a729-a5ae98c41cf9] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 752.542568] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-282d8ab0-6bbf-4e7e-911c-8810a7f56d42 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.564030] env[63531]: DEBUG oslo_vmware.api [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]520631b0-7dd3-f4b5-534f-587f274c4eee, 'name': SearchDatastore_Task, 'duration_secs': 0.024275} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.564270] env[63531]: DEBUG oslo_concurrency.lockutils [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 752.564726] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 752.566631] env[63531]: DEBUG oslo_concurrency.lockutils [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 752.566631] env[63531]: DEBUG oslo_concurrency.lockutils [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.566631] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 752.566631] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b2f48aa6-6de7-4110-99ae-20d5dab979c6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.577226] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 752.577226] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 752.577226] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4335a1e8-a570-44d4-a5e0-4f1bb67cebe2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.582014] env[63531]: DEBUG oslo_vmware.api [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Waiting for the task: (returnval){ [ 752.582014] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52422a39-12eb-84fd-111d-4a54471b4105" [ 752.582014] env[63531]: _type = "Task" [ 752.582014] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.601472] env[63531]: DEBUG oslo_vmware.api [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52422a39-12eb-84fd-111d-4a54471b4105, 'name': SearchDatastore_Task, 'duration_secs': 0.00826} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.601472] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c1887ee3-d617-4de3-b98a-b0bedd40b449 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.608451] env[63531]: DEBUG oslo_concurrency.lockutils [None req-524a4683-a8e1-4d4b-87d9-f141433baa74 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.061s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 752.610760] env[63531]: DEBUG oslo_vmware.api [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Waiting for the task: (returnval){ [ 752.610760] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52165076-3501-484d-39bb-bd8c36412dd8" [ 752.610760] env[63531]: _type = "Task" [ 752.610760] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.611201] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.868s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 752.613042] env[63531]: INFO nova.compute.claims [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 752.625215] env[63531]: DEBUG oslo_vmware.api [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52165076-3501-484d-39bb-bd8c36412dd8, 'name': SearchDatastore_Task, 'duration_secs': 0.008722} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.625215] env[63531]: DEBUG oslo_concurrency.lockutils [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 752.625215] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74/e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 752.625381] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7faea477-41c8-4715-8b4f-762a7a2f08ef {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.633849] env[63531]: DEBUG oslo_vmware.api [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Waiting for the task: (returnval){ [ 752.633849] env[63531]: value = "task-1117668" [ 752.633849] env[63531]: _type = "Task" [ 752.633849] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.642026] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a56197e5-b9a2-4ba5-a23f-e676b6cc2b13 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] [instance: 0af99240-fa7a-4eed-a729-a5ae98c41cf9] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 752.642026] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a56197e5-b9a2-4ba5-a23f-e676b6cc2b13 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] [instance: 0af99240-fa7a-4eed-a729-a5ae98c41cf9] Deleting contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 752.642026] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-a56197e5-b9a2-4ba5-a23f-e676b6cc2b13 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Deleting the datastore file [datastore2] 0af99240-fa7a-4eed-a729-a5ae98c41cf9 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 752.642026] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-96db81eb-a782-4b04-9cba-120e244c52a1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.650825] env[63531]: DEBUG oslo_vmware.api [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Task: {'id': task-1117668, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.653381] env[63531]: INFO nova.scheduler.client.report [None req-524a4683-a8e1-4d4b-87d9-f141433baa74 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Deleted allocations for instance f0276e34-1de3-482a-82c7-1439d40fd85e [ 752.656228] env[63531]: DEBUG oslo_vmware.api [None req-a56197e5-b9a2-4ba5-a23f-e676b6cc2b13 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Waiting for the task: (returnval){ [ 752.656228] env[63531]: value = "task-1117669" [ 752.656228] env[63531]: _type = "Task" [ 752.656228] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.669202] env[63531]: DEBUG oslo_vmware.api [None req-a56197e5-b9a2-4ba5-a23f-e676b6cc2b13 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Task: {'id': task-1117669, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.690961] env[63531]: DEBUG oslo_vmware.api [None req-9b7b17b4-8f5f-4f77-83b4-2a61816669ba tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117666, 'name': ReconfigVM_Task, 'duration_secs': 0.308171} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.691265] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-9b7b17b4-8f5f-4f77-83b4-2a61816669ba tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Reconfigured VM instance instance-0000002b to attach disk [datastore2] e7fb00a3-0d2e-4f54-950d-337307112d7a/9d550399-588c-4bd5-8436-eba9b523bd13-rescue.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 752.692142] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45222ee6-6033-4c50-92e7-22aeab5499ae {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.726519] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-34f8f73a-1f03-4c8a-881c-a73099e833cf {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.743553] env[63531]: DEBUG oslo_vmware.api [None req-9b7b17b4-8f5f-4f77-83b4-2a61816669ba tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Waiting for the task: (returnval){ [ 752.743553] env[63531]: value = "task-1117670" [ 752.743553] env[63531]: _type = "Task" [ 752.743553] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.755049] env[63531]: DEBUG oslo_vmware.api [None req-9b7b17b4-8f5f-4f77-83b4-2a61816669ba tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117670, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.755049] env[63531]: DEBUG nova.network.neutron [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Successfully updated port: 4c96ebbd-98fa-4d12-a388-b9d270b7aaae {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 752.879613] env[63531]: DEBUG oslo_concurrency.lockutils [req-72f3e474-a36d-4868-92f9-edc6560f0576 req-6c41cd46-3c8d-4d8f-8d14-90c8235e08d1 service nova] Releasing lock "refresh_cache-e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 753.155344] env[63531]: DEBUG oslo_vmware.api [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Task: {'id': task-1117668, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.176749] env[63531]: DEBUG oslo_vmware.api [None req-a56197e5-b9a2-4ba5-a23f-e676b6cc2b13 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Task: {'id': task-1117669, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.216427} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.177508] env[63531]: DEBUG oslo_concurrency.lockutils [None req-524a4683-a8e1-4d4b-87d9-f141433baa74 tempest-AttachInterfacesV270Test-1742456677 tempest-AttachInterfacesV270Test-1742456677-project-member] Lock "f0276e34-1de3-482a-82c7-1439d40fd85e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.967s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 753.178850] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-a56197e5-b9a2-4ba5-a23f-e676b6cc2b13 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 753.179144] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a56197e5-b9a2-4ba5-a23f-e676b6cc2b13 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] [instance: 0af99240-fa7a-4eed-a729-a5ae98c41cf9] Deleted contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 753.179354] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a56197e5-b9a2-4ba5-a23f-e676b6cc2b13 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] [instance: 0af99240-fa7a-4eed-a729-a5ae98c41cf9] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 753.179762] env[63531]: INFO nova.compute.manager [None req-a56197e5-b9a2-4ba5-a23f-e676b6cc2b13 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] [instance: 0af99240-fa7a-4eed-a729-a5ae98c41cf9] Took 1.17 seconds to destroy the instance on the hypervisor. [ 753.180131] env[63531]: DEBUG oslo.service.loopingcall [None req-a56197e5-b9a2-4ba5-a23f-e676b6cc2b13 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 753.180436] env[63531]: DEBUG nova.compute.manager [-] [instance: 0af99240-fa7a-4eed-a729-a5ae98c41cf9] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 753.180556] env[63531]: DEBUG nova.network.neutron [-] [instance: 0af99240-fa7a-4eed-a729-a5ae98c41cf9] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 753.256574] env[63531]: DEBUG oslo_vmware.api [None req-9b7b17b4-8f5f-4f77-83b4-2a61816669ba tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117670, 'name': ReconfigVM_Task, 'duration_secs': 0.204613} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.256990] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b7b17b4-8f5f-4f77-83b4-2a61816669ba tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 753.259413] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-48e08413-9498-4bac-ae77-1726ffa512b5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.266809] env[63531]: DEBUG oslo_vmware.api [None req-9b7b17b4-8f5f-4f77-83b4-2a61816669ba tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Waiting for the task: (returnval){ [ 753.266809] env[63531]: value = "task-1117671" [ 753.266809] env[63531]: _type = "Task" [ 753.266809] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.276216] env[63531]: DEBUG oslo_vmware.api [None req-9b7b17b4-8f5f-4f77-83b4-2a61816669ba tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117671, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.644697] env[63531]: DEBUG oslo_vmware.api [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Task: {'id': task-1117668, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.57413} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.647574] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74/e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 753.647816] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 753.648277] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-09970cf5-c08e-4e14-8448-8eba66c81f12 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.659328] env[63531]: DEBUG oslo_vmware.api [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Waiting for the task: (returnval){ [ 753.659328] env[63531]: value = "task-1117672" [ 753.659328] env[63531]: _type = "Task" [ 753.659328] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.672362] env[63531]: DEBUG oslo_vmware.api [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Task: {'id': task-1117672, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.776909] env[63531]: DEBUG oslo_vmware.api [None req-9b7b17b4-8f5f-4f77-83b4-2a61816669ba tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117671, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.975643] env[63531]: DEBUG nova.network.neutron [-] [instance: 0af99240-fa7a-4eed-a729-a5ae98c41cf9] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.009045] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Acquiring lock "eb6c84d7-e0e6-4999-9018-889b99dc67ad" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 754.009045] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Lock "eb6c84d7-e0e6-4999-9018-889b99dc67ad" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 754.030750] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a5843b3-6d5e-4528-a2a5-7c791be3b847 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.038600] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d28e7a2-35df-4017-b690-3ea574153a28 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.069975] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5882d9c2-72fa-41af-a0f4-c762034aacb1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.077446] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b09cf624-5de4-4908-a74d-c89fa8ea21e7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.092786] env[63531]: DEBUG nova.compute.provider_tree [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Updating inventory in ProviderTree for provider 9996a314-a62e-4282-8252-f387e7e3fabb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 754.172948] env[63531]: DEBUG oslo_vmware.api [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Task: {'id': task-1117672, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065698} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.172948] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 754.172948] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bd6fab4-ca34-4c34-a284-797b860c2175 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.197086] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74] Reconfiguring VM instance instance-0000002d to attach disk [datastore1] e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74/e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 754.197430] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cfc71a5d-180c-4cc0-bc55-e5d83401340c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.217227] env[63531]: DEBUG oslo_vmware.api [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Waiting for the task: (returnval){ [ 754.217227] env[63531]: value = "task-1117673" [ 754.217227] env[63531]: _type = "Task" [ 754.217227] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.225607] env[63531]: DEBUG oslo_vmware.api [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Task: {'id': task-1117673, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.281194] env[63531]: DEBUG oslo_vmware.api [None req-9b7b17b4-8f5f-4f77-83b4-2a61816669ba tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117671, 'name': PowerOnVM_Task, 'duration_secs': 0.990489} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.281306] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b7b17b4-8f5f-4f77-83b4-2a61816669ba tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 754.283970] env[63531]: DEBUG nova.compute.manager [None req-9b7b17b4-8f5f-4f77-83b4-2a61816669ba tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 754.284770] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88ad709b-a1f7-4e57-8795-fa35942ce79e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.477323] env[63531]: INFO nova.compute.manager [-] [instance: 0af99240-fa7a-4eed-a729-a5ae98c41cf9] Took 1.30 seconds to deallocate network for instance. [ 754.617153] env[63531]: ERROR nova.scheduler.client.report [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [req-b7bb69fa-5605-4558-bc89-67550751df7d] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 9996a314-a62e-4282-8252-f387e7e3fabb. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-b7bb69fa-5605-4558-bc89-67550751df7d"}]} [ 754.639901] env[63531]: DEBUG nova.scheduler.client.report [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Refreshing inventories for resource provider 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 754.647822] env[63531]: DEBUG nova.compute.manager [req-6cfd18f5-abb2-4c7d-a754-1f5652cb5886 req-c75260fa-657e-45e2-99ec-ec555992d017 service nova] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Received event network-changed-91733d5a-84d6-48df-ba4d-7e362a642282 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 754.647822] env[63531]: DEBUG nova.compute.manager [req-6cfd18f5-abb2-4c7d-a754-1f5652cb5886 req-c75260fa-657e-45e2-99ec-ec555992d017 service nova] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Refreshing instance network info cache due to event network-changed-91733d5a-84d6-48df-ba4d-7e362a642282. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 754.648093] env[63531]: DEBUG oslo_concurrency.lockutils [req-6cfd18f5-abb2-4c7d-a754-1f5652cb5886 req-c75260fa-657e-45e2-99ec-ec555992d017 service nova] Acquiring lock "refresh_cache-c3920e60-b27f-496b-b616-afb23ab0bc26" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 754.648093] env[63531]: DEBUG oslo_concurrency.lockutils [req-6cfd18f5-abb2-4c7d-a754-1f5652cb5886 req-c75260fa-657e-45e2-99ec-ec555992d017 service nova] Acquired lock "refresh_cache-c3920e60-b27f-496b-b616-afb23ab0bc26" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 754.648267] env[63531]: DEBUG nova.network.neutron [req-6cfd18f5-abb2-4c7d-a754-1f5652cb5886 req-c75260fa-657e-45e2-99ec-ec555992d017 service nova] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Refreshing network info cache for port 91733d5a-84d6-48df-ba4d-7e362a642282 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 754.661232] env[63531]: DEBUG nova.scheduler.client.report [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Updating ProviderTree inventory for provider 9996a314-a62e-4282-8252-f387e7e3fabb from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 754.661350] env[63531]: DEBUG nova.compute.provider_tree [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Updating inventory in ProviderTree for provider 9996a314-a62e-4282-8252-f387e7e3fabb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 754.675941] env[63531]: DEBUG nova.scheduler.client.report [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Refreshing aggregate associations for resource provider 9996a314-a62e-4282-8252-f387e7e3fabb, aggregates: None {{(pid=63531) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 754.693520] env[63531]: DEBUG nova.scheduler.client.report [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Refreshing trait associations for resource provider 9996a314-a62e-4282-8252-f387e7e3fabb, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE {{(pid=63531) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 754.729509] env[63531]: DEBUG oslo_vmware.api [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Task: {'id': task-1117673, 'name': ReconfigVM_Task, 'duration_secs': 0.280579} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.729836] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74] Reconfigured VM instance instance-0000002d to attach disk [datastore1] e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74/e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 754.730644] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c5b0cd05-062c-407e-901a-d3905e599eb9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.736988] env[63531]: DEBUG oslo_vmware.api [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Waiting for the task: (returnval){ [ 754.736988] env[63531]: value = "task-1117674" [ 754.736988] env[63531]: _type = "Task" [ 754.736988] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.746150] env[63531]: DEBUG oslo_vmware.api [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Task: {'id': task-1117674, 'name': Rename_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.907741] env[63531]: DEBUG nova.compute.manager [req-3d569159-0336-4d2e-98e1-e8782ab0e50d req-a63a4161-5da6-4ef5-adae-b50d38fb8388 service nova] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Received event network-vif-plugged-4c96ebbd-98fa-4d12-a388-b9d270b7aaae {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 754.907991] env[63531]: DEBUG oslo_concurrency.lockutils [req-3d569159-0336-4d2e-98e1-e8782ab0e50d req-a63a4161-5da6-4ef5-adae-b50d38fb8388 service nova] Acquiring lock "e1693594-d8bb-4002-b71c-7013efb81d16-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 754.908222] env[63531]: DEBUG oslo_concurrency.lockutils [req-3d569159-0336-4d2e-98e1-e8782ab0e50d req-a63a4161-5da6-4ef5-adae-b50d38fb8388 service nova] Lock "e1693594-d8bb-4002-b71c-7013efb81d16-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 754.908388] env[63531]: DEBUG oslo_concurrency.lockutils [req-3d569159-0336-4d2e-98e1-e8782ab0e50d req-a63a4161-5da6-4ef5-adae-b50d38fb8388 service nova] Lock "e1693594-d8bb-4002-b71c-7013efb81d16-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 754.909266] env[63531]: DEBUG nova.compute.manager [req-3d569159-0336-4d2e-98e1-e8782ab0e50d req-a63a4161-5da6-4ef5-adae-b50d38fb8388 service nova] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] No waiting events found dispatching network-vif-plugged-4c96ebbd-98fa-4d12-a388-b9d270b7aaae {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 754.909266] env[63531]: WARNING nova.compute.manager [req-3d569159-0336-4d2e-98e1-e8782ab0e50d req-a63a4161-5da6-4ef5-adae-b50d38fb8388 service nova] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Received unexpected event network-vif-plugged-4c96ebbd-98fa-4d12-a388-b9d270b7aaae for instance with vm_state building and task_state spawning. [ 754.909266] env[63531]: DEBUG nova.compute.manager [req-3d569159-0336-4d2e-98e1-e8782ab0e50d req-a63a4161-5da6-4ef5-adae-b50d38fb8388 service nova] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Received event network-changed-4c96ebbd-98fa-4d12-a388-b9d270b7aaae {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 754.909266] env[63531]: DEBUG nova.compute.manager [req-3d569159-0336-4d2e-98e1-e8782ab0e50d req-a63a4161-5da6-4ef5-adae-b50d38fb8388 service nova] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Refreshing instance network info cache due to event network-changed-4c96ebbd-98fa-4d12-a388-b9d270b7aaae. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 754.909718] env[63531]: DEBUG oslo_concurrency.lockutils [req-3d569159-0336-4d2e-98e1-e8782ab0e50d req-a63a4161-5da6-4ef5-adae-b50d38fb8388 service nova] Acquiring lock "refresh_cache-e1693594-d8bb-4002-b71c-7013efb81d16" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 754.909868] env[63531]: DEBUG oslo_concurrency.lockutils [req-3d569159-0336-4d2e-98e1-e8782ab0e50d req-a63a4161-5da6-4ef5-adae-b50d38fb8388 service nova] Acquired lock "refresh_cache-e1693594-d8bb-4002-b71c-7013efb81d16" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 754.910047] env[63531]: DEBUG nova.network.neutron [req-3d569159-0336-4d2e-98e1-e8782ab0e50d req-a63a4161-5da6-4ef5-adae-b50d38fb8388 service nova] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Refreshing network info cache for port 4c96ebbd-98fa-4d12-a388-b9d270b7aaae {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 754.986337] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a56197e5-b9a2-4ba5-a23f-e676b6cc2b13 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 755.098645] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19ef2118-7617-4644-876a-a235832233a3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.108971] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac3ee545-9ad8-40bf-a740-792e24ffe573 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.146669] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a76acad-d963-4fe9-a616-05d983938912 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.158084] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2c356a8-0cc9-4af7-85b9-9646db0ee3a7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.175067] env[63531]: DEBUG nova.compute.provider_tree [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Updating inventory in ProviderTree for provider 9996a314-a62e-4282-8252-f387e7e3fabb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 755.219196] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Acquiring lock "f46d8a85-6ab7-436a-bc5e-40d0f80974e5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 755.219725] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lock "f46d8a85-6ab7-436a-bc5e-40d0f80974e5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 755.250273] env[63531]: DEBUG oslo_vmware.api [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Task: {'id': task-1117674, 'name': Rename_Task, 'duration_secs': 0.142965} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.250546] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 755.250781] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8acfc942-4f0b-4c77-beab-01bac30487b2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.257288] env[63531]: DEBUG oslo_vmware.api [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Waiting for the task: (returnval){ [ 755.257288] env[63531]: value = "task-1117675" [ 755.257288] env[63531]: _type = "Task" [ 755.257288] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.265481] env[63531]: DEBUG oslo_vmware.api [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Task: {'id': task-1117675, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.512921] env[63531]: DEBUG nova.network.neutron [req-3d569159-0336-4d2e-98e1-e8782ab0e50d req-a63a4161-5da6-4ef5-adae-b50d38fb8388 service nova] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 755.639559] env[63531]: DEBUG nova.network.neutron [req-6cfd18f5-abb2-4c7d-a754-1f5652cb5886 req-c75260fa-657e-45e2-99ec-ec555992d017 service nova] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Updated VIF entry in instance network info cache for port 91733d5a-84d6-48df-ba4d-7e362a642282. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 755.639953] env[63531]: DEBUG nova.network.neutron [req-6cfd18f5-abb2-4c7d-a754-1f5652cb5886 req-c75260fa-657e-45e2-99ec-ec555992d017 service nova] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Updating instance_info_cache with network_info: [{"id": "91733d5a-84d6-48df-ba4d-7e362a642282", "address": "fa:16:3e:df:a1:31", "network": {"id": "d28b2087-329c-4917-aa6e-83ab83a63a74", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1054369059-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.158", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a634a61020b94d62a8fb37a3f96acc10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91733d5a-84", "ovs_interfaceid": "91733d5a-84d6-48df-ba4d-7e362a642282", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 755.703947] env[63531]: ERROR nova.scheduler.client.report [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [req-cbb0200d-92c5-4e7b-b061-d3855e368d33] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 9996a314-a62e-4282-8252-f387e7e3fabb. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-cbb0200d-92c5-4e7b-b061-d3855e368d33"}]} [ 755.735883] env[63531]: DEBUG nova.scheduler.client.report [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Refreshing inventories for resource provider 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 755.755724] env[63531]: DEBUG nova.scheduler.client.report [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Updating ProviderTree inventory for provider 9996a314-a62e-4282-8252-f387e7e3fabb from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 755.755951] env[63531]: DEBUG nova.compute.provider_tree [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Updating inventory in ProviderTree for provider 9996a314-a62e-4282-8252-f387e7e3fabb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 755.771800] env[63531]: DEBUG oslo_vmware.api [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Task: {'id': task-1117675, 'name': PowerOnVM_Task, 'duration_secs': 0.491301} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.772732] env[63531]: DEBUG nova.scheduler.client.report [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Refreshing aggregate associations for resource provider 9996a314-a62e-4282-8252-f387e7e3fabb, aggregates: None {{(pid=63531) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 755.775503] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 755.775503] env[63531]: INFO nova.compute.manager [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74] Took 7.80 seconds to spawn the instance on the hypervisor. [ 755.775637] env[63531]: DEBUG nova.compute.manager [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 755.776564] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7309e778-d46c-4896-895a-c2d6dcde17db {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.793460] env[63531]: DEBUG nova.scheduler.client.report [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Refreshing trait associations for resource provider 9996a314-a62e-4282-8252-f387e7e3fabb, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE {{(pid=63531) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 755.844736] env[63531]: DEBUG nova.network.neutron [req-3d569159-0336-4d2e-98e1-e8782ab0e50d req-a63a4161-5da6-4ef5-adae-b50d38fb8388 service nova] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 756.143367] env[63531]: DEBUG oslo_concurrency.lockutils [req-6cfd18f5-abb2-4c7d-a754-1f5652cb5886 req-c75260fa-657e-45e2-99ec-ec555992d017 service nova] Releasing lock "refresh_cache-c3920e60-b27f-496b-b616-afb23ab0bc26" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 756.238299] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82b8c34e-9cff-49eb-850d-030599d37491 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.251019] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53540184-14db-42f0-ae52-9c43b1ecef02 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.280019] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-268fc8f8-f00e-4b16-977e-daa8053b984a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.292032] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb4a72a7-7063-4944-87d1-9f9cdcb16500 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.307228] env[63531]: DEBUG nova.compute.provider_tree [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Updating inventory in ProviderTree for provider 9996a314-a62e-4282-8252-f387e7e3fabb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 756.313165] env[63531]: INFO nova.compute.manager [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74] Took 29.24 seconds to build instance. [ 756.347587] env[63531]: DEBUG oslo_concurrency.lockutils [req-3d569159-0336-4d2e-98e1-e8782ab0e50d req-a63a4161-5da6-4ef5-adae-b50d38fb8388 service nova] Releasing lock "refresh_cache-e1693594-d8bb-4002-b71c-7013efb81d16" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 756.551084] env[63531]: DEBUG nova.network.neutron [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Successfully updated port: 5765492b-2d11-438b-b9a4-7514ce2ac73e {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 756.818549] env[63531]: DEBUG oslo_concurrency.lockutils [None req-58b04e6d-36e4-468b-bf09-04bb73aa199c tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Lock "e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 130.898s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 756.855303] env[63531]: DEBUG nova.scheduler.client.report [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Updated inventory for provider 9996a314-a62e-4282-8252-f387e7e3fabb with generation 68 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 756.856276] env[63531]: DEBUG nova.compute.provider_tree [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Updating resource provider 9996a314-a62e-4282-8252-f387e7e3fabb generation from 68 to 69 during operation: update_inventory {{(pid=63531) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 756.856276] env[63531]: DEBUG nova.compute.provider_tree [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Updating inventory in ProviderTree for provider 9996a314-a62e-4282-8252-f387e7e3fabb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 757.324349] env[63531]: DEBUG nova.compute.manager [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 757.364205] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 4.750s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 757.364205] env[63531]: DEBUG nova.compute.manager [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 757.366982] env[63531]: DEBUG oslo_concurrency.lockutils [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.096s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 757.370368] env[63531]: INFO nova.compute.claims [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] [instance: dccb2335-4220-4570-861b-16ff2d8f5a85] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 757.746890] env[63531]: DEBUG nova.compute.manager [req-213a90f9-e0e5-4a30-9503-30c51843a523 req-a9e40dce-3976-4117-a1e2-f3e578ef0dbd service nova] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Received event network-changed-91733d5a-84d6-48df-ba4d-7e362a642282 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 757.747053] env[63531]: DEBUG nova.compute.manager [req-213a90f9-e0e5-4a30-9503-30c51843a523 req-a9e40dce-3976-4117-a1e2-f3e578ef0dbd service nova] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Refreshing instance network info cache due to event network-changed-91733d5a-84d6-48df-ba4d-7e362a642282. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 757.747311] env[63531]: DEBUG oslo_concurrency.lockutils [req-213a90f9-e0e5-4a30-9503-30c51843a523 req-a9e40dce-3976-4117-a1e2-f3e578ef0dbd service nova] Acquiring lock "refresh_cache-c3920e60-b27f-496b-b616-afb23ab0bc26" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 757.747420] env[63531]: DEBUG oslo_concurrency.lockutils [req-213a90f9-e0e5-4a30-9503-30c51843a523 req-a9e40dce-3976-4117-a1e2-f3e578ef0dbd service nova] Acquired lock "refresh_cache-c3920e60-b27f-496b-b616-afb23ab0bc26" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.747550] env[63531]: DEBUG nova.network.neutron [req-213a90f9-e0e5-4a30-9503-30c51843a523 req-a9e40dce-3976-4117-a1e2-f3e578ef0dbd service nova] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Refreshing network info cache for port 91733d5a-84d6-48df-ba4d-7e362a642282 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 757.765142] env[63531]: DEBUG nova.compute.manager [req-d3035e38-7f43-4b86-9e49-4cffbd7152e6 req-e6eaad17-95f7-416d-8a07-aeb3e8582a29 service nova] [instance: 0af99240-fa7a-4eed-a729-a5ae98c41cf9] Received event network-vif-deleted-c2356641-c89b-4610-ade8-8cc4e392021b {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 757.855069] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 757.868820] env[63531]: DEBUG nova.compute.utils [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 757.870272] env[63531]: DEBUG nova.compute.manager [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 757.871488] env[63531]: DEBUG nova.network.neutron [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 757.957099] env[63531]: DEBUG nova.policy [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '597d424341124b7db23dc7a104107148', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '308eabafea9c4188a58a0f1c22074d2f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 757.999641] env[63531]: DEBUG oslo_concurrency.lockutils [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Acquiring lock "e0074e2e-e8bd-4bcb-9ad9-271a10e079e4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 757.999898] env[63531]: DEBUG oslo_concurrency.lockutils [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Lock "e0074e2e-e8bd-4bcb-9ad9-271a10e079e4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 758.377043] env[63531]: DEBUG nova.compute.manager [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 758.576131] env[63531]: DEBUG nova.network.neutron [req-213a90f9-e0e5-4a30-9503-30c51843a523 req-a9e40dce-3976-4117-a1e2-f3e578ef0dbd service nova] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Updated VIF entry in instance network info cache for port 91733d5a-84d6-48df-ba4d-7e362a642282. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 758.576675] env[63531]: DEBUG nova.network.neutron [req-213a90f9-e0e5-4a30-9503-30c51843a523 req-a9e40dce-3976-4117-a1e2-f3e578ef0dbd service nova] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Updating instance_info_cache with network_info: [{"id": "91733d5a-84d6-48df-ba4d-7e362a642282", "address": "fa:16:3e:df:a1:31", "network": {"id": "d28b2087-329c-4917-aa6e-83ab83a63a74", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1054369059-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a634a61020b94d62a8fb37a3f96acc10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91733d5a-84", "ovs_interfaceid": "91733d5a-84d6-48df-ba4d-7e362a642282", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.618568] env[63531]: DEBUG nova.network.neutron [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Successfully created port: 262abfb7-7bcf-40e6-9ce3-b502de45b3b4 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 758.838058] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-907819f0-4927-4df7-881c-f4b3fd6df8e9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.845949] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27541803-2fd0-458d-b591-60916f8d01c1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.879282] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f793109c-d947-4b61-8a46-ee4102ddddf8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.893025] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fa91c33-e6db-4ed7-a73c-ef1839078de6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.904147] env[63531]: DEBUG nova.compute.provider_tree [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 759.081749] env[63531]: DEBUG oslo_concurrency.lockutils [req-213a90f9-e0e5-4a30-9503-30c51843a523 req-a9e40dce-3976-4117-a1e2-f3e578ef0dbd service nova] Releasing lock "refresh_cache-c3920e60-b27f-496b-b616-afb23ab0bc26" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 759.082067] env[63531]: DEBUG nova.compute.manager [req-213a90f9-e0e5-4a30-9503-30c51843a523 req-a9e40dce-3976-4117-a1e2-f3e578ef0dbd service nova] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Received event network-vif-plugged-5765492b-2d11-438b-b9a4-7514ce2ac73e {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 759.082276] env[63531]: DEBUG oslo_concurrency.lockutils [req-213a90f9-e0e5-4a30-9503-30c51843a523 req-a9e40dce-3976-4117-a1e2-f3e578ef0dbd service nova] Acquiring lock "e1693594-d8bb-4002-b71c-7013efb81d16-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 759.082488] env[63531]: DEBUG oslo_concurrency.lockutils [req-213a90f9-e0e5-4a30-9503-30c51843a523 req-a9e40dce-3976-4117-a1e2-f3e578ef0dbd service nova] Lock "e1693594-d8bb-4002-b71c-7013efb81d16-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 759.082642] env[63531]: DEBUG oslo_concurrency.lockutils [req-213a90f9-e0e5-4a30-9503-30c51843a523 req-a9e40dce-3976-4117-a1e2-f3e578ef0dbd service nova] Lock "e1693594-d8bb-4002-b71c-7013efb81d16-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 759.082801] env[63531]: DEBUG nova.compute.manager [req-213a90f9-e0e5-4a30-9503-30c51843a523 req-a9e40dce-3976-4117-a1e2-f3e578ef0dbd service nova] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] No waiting events found dispatching network-vif-plugged-5765492b-2d11-438b-b9a4-7514ce2ac73e {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 759.082966] env[63531]: WARNING nova.compute.manager [req-213a90f9-e0e5-4a30-9503-30c51843a523 req-a9e40dce-3976-4117-a1e2-f3e578ef0dbd service nova] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Received unexpected event network-vif-plugged-5765492b-2d11-438b-b9a4-7514ce2ac73e for instance with vm_state building and task_state spawning. [ 759.083138] env[63531]: DEBUG nova.compute.manager [req-213a90f9-e0e5-4a30-9503-30c51843a523 req-a9e40dce-3976-4117-a1e2-f3e578ef0dbd service nova] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Received event network-changed-5765492b-2d11-438b-b9a4-7514ce2ac73e {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 759.083406] env[63531]: DEBUG nova.compute.manager [req-213a90f9-e0e5-4a30-9503-30c51843a523 req-a9e40dce-3976-4117-a1e2-f3e578ef0dbd service nova] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Refreshing instance network info cache due to event network-changed-5765492b-2d11-438b-b9a4-7514ce2ac73e. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 759.083473] env[63531]: DEBUG oslo_concurrency.lockutils [req-213a90f9-e0e5-4a30-9503-30c51843a523 req-a9e40dce-3976-4117-a1e2-f3e578ef0dbd service nova] Acquiring lock "refresh_cache-e1693594-d8bb-4002-b71c-7013efb81d16" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 759.083587] env[63531]: DEBUG oslo_concurrency.lockutils [req-213a90f9-e0e5-4a30-9503-30c51843a523 req-a9e40dce-3976-4117-a1e2-f3e578ef0dbd service nova] Acquired lock "refresh_cache-e1693594-d8bb-4002-b71c-7013efb81d16" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 759.083735] env[63531]: DEBUG nova.network.neutron [req-213a90f9-e0e5-4a30-9503-30c51843a523 req-a9e40dce-3976-4117-a1e2-f3e578ef0dbd service nova] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Refreshing network info cache for port 5765492b-2d11-438b-b9a4-7514ce2ac73e {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 759.397176] env[63531]: DEBUG nova.compute.manager [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 759.407347] env[63531]: DEBUG nova.scheduler.client.report [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 759.433216] env[63531]: DEBUG nova.virt.hardware [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 759.433749] env[63531]: DEBUG nova.virt.hardware [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 759.433749] env[63531]: DEBUG nova.virt.hardware [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 759.433980] env[63531]: DEBUG nova.virt.hardware [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 759.433980] env[63531]: DEBUG nova.virt.hardware [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 759.434882] env[63531]: DEBUG nova.virt.hardware [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 759.435652] env[63531]: DEBUG nova.virt.hardware [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 759.435911] env[63531]: DEBUG nova.virt.hardware [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 759.436168] env[63531]: DEBUG nova.virt.hardware [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 759.436489] env[63531]: DEBUG nova.virt.hardware [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 759.436583] env[63531]: DEBUG nova.virt.hardware [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 759.437736] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b86b512-b3f6-4bd9-a8d0-c9cc150b7c9c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.447192] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54decf8c-353a-411d-a3d0-265cb52f12b3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.621641] env[63531]: DEBUG nova.network.neutron [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Successfully updated port: 4da738f4-b2e9-40c4-9a02-43b6ed658cfa {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 759.634304] env[63531]: DEBUG nova.network.neutron [req-213a90f9-e0e5-4a30-9503-30c51843a523 req-a9e40dce-3976-4117-a1e2-f3e578ef0dbd service nova] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 759.777712] env[63531]: DEBUG nova.network.neutron [req-213a90f9-e0e5-4a30-9503-30c51843a523 req-a9e40dce-3976-4117-a1e2-f3e578ef0dbd service nova] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.913162] env[63531]: DEBUG oslo_concurrency.lockutils [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.546s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 759.913702] env[63531]: DEBUG nova.compute.manager [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] [instance: dccb2335-4220-4570-861b-16ff2d8f5a85] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 759.917090] env[63531]: DEBUG oslo_concurrency.lockutils [None req-db4fd4eb-d9b5-45f3-8b8c-2517c717ed2c tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.136s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 759.917937] env[63531]: DEBUG nova.objects.instance [None req-db4fd4eb-d9b5-45f3-8b8c-2517c717ed2c tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Lazy-loading 'resources' on Instance uuid c3d597b4-1dc3-48d8-9bee-e73c8929181b {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 760.043175] env[63531]: DEBUG nova.compute.manager [req-b11ca775-6519-40dc-8cc7-d400463b9865 req-b932a322-7c03-4cd1-815c-bfce8f79e0bb service nova] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Received event network-changed-91733d5a-84d6-48df-ba4d-7e362a642282 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 760.043398] env[63531]: DEBUG nova.compute.manager [req-b11ca775-6519-40dc-8cc7-d400463b9865 req-b932a322-7c03-4cd1-815c-bfce8f79e0bb service nova] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Refreshing instance network info cache due to event network-changed-91733d5a-84d6-48df-ba4d-7e362a642282. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 760.043613] env[63531]: DEBUG oslo_concurrency.lockutils [req-b11ca775-6519-40dc-8cc7-d400463b9865 req-b932a322-7c03-4cd1-815c-bfce8f79e0bb service nova] Acquiring lock "refresh_cache-c3920e60-b27f-496b-b616-afb23ab0bc26" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 760.043753] env[63531]: DEBUG oslo_concurrency.lockutils [req-b11ca775-6519-40dc-8cc7-d400463b9865 req-b932a322-7c03-4cd1-815c-bfce8f79e0bb service nova] Acquired lock "refresh_cache-c3920e60-b27f-496b-b616-afb23ab0bc26" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 760.043910] env[63531]: DEBUG nova.network.neutron [req-b11ca775-6519-40dc-8cc7-d400463b9865 req-b932a322-7c03-4cd1-815c-bfce8f79e0bb service nova] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Refreshing network info cache for port 91733d5a-84d6-48df-ba4d-7e362a642282 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 760.125684] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Acquiring lock "refresh_cache-e1693594-d8bb-4002-b71c-7013efb81d16" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 760.282675] env[63531]: DEBUG oslo_concurrency.lockutils [req-213a90f9-e0e5-4a30-9503-30c51843a523 req-a9e40dce-3976-4117-a1e2-f3e578ef0dbd service nova] Releasing lock "refresh_cache-e1693594-d8bb-4002-b71c-7013efb81d16" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 760.283103] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Acquired lock "refresh_cache-e1693594-d8bb-4002-b71c-7013efb81d16" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 760.283263] env[63531]: DEBUG nova.network.neutron [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 760.421909] env[63531]: DEBUG nova.compute.utils [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 760.429886] env[63531]: DEBUG nova.compute.manager [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] [instance: dccb2335-4220-4570-861b-16ff2d8f5a85] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 760.429886] env[63531]: DEBUG nova.network.neutron [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] [instance: dccb2335-4220-4570-861b-16ff2d8f5a85] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 760.438602] env[63531]: DEBUG oslo_concurrency.lockutils [None req-785bdba5-357d-4d51-af3f-e636dea49e89 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Acquiring lock "c3920e60-b27f-496b-b616-afb23ab0bc26" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.438866] env[63531]: DEBUG oslo_concurrency.lockutils [None req-785bdba5-357d-4d51-af3f-e636dea49e89 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Lock "c3920e60-b27f-496b-b616-afb23ab0bc26" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 760.439088] env[63531]: DEBUG oslo_concurrency.lockutils [None req-785bdba5-357d-4d51-af3f-e636dea49e89 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Acquiring lock "c3920e60-b27f-496b-b616-afb23ab0bc26-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.439361] env[63531]: DEBUG oslo_concurrency.lockutils [None req-785bdba5-357d-4d51-af3f-e636dea49e89 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Lock "c3920e60-b27f-496b-b616-afb23ab0bc26-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 760.439537] env[63531]: DEBUG oslo_concurrency.lockutils [None req-785bdba5-357d-4d51-af3f-e636dea49e89 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Lock "c3920e60-b27f-496b-b616-afb23ab0bc26-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 760.442059] env[63531]: INFO nova.compute.manager [None req-785bdba5-357d-4d51-af3f-e636dea49e89 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Terminating instance [ 760.444279] env[63531]: DEBUG nova.compute.manager [None req-785bdba5-357d-4d51-af3f-e636dea49e89 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 760.444279] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-785bdba5-357d-4d51-af3f-e636dea49e89 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 760.444975] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e64675f-681b-4368-a337-88d40c2356fd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.458185] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-785bdba5-357d-4d51-af3f-e636dea49e89 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 760.458443] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-54e4ede0-7a35-4a38-be36-5e8357010c69 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.466122] env[63531]: DEBUG oslo_vmware.api [None req-785bdba5-357d-4d51-af3f-e636dea49e89 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Waiting for the task: (returnval){ [ 760.466122] env[63531]: value = "task-1117676" [ 760.466122] env[63531]: _type = "Task" [ 760.466122] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.475618] env[63531]: DEBUG oslo_vmware.api [None req-785bdba5-357d-4d51-af3f-e636dea49e89 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Task: {'id': task-1117676, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.478460] env[63531]: DEBUG nova.policy [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dd645d72257649fba0a9afb2f320221c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c4d75d29af5145bfb5140aabaeabf7e0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 760.847121] env[63531]: DEBUG nova.network.neutron [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 760.887754] env[63531]: DEBUG nova.network.neutron [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] [instance: dccb2335-4220-4570-861b-16ff2d8f5a85] Successfully created port: d69ccbf2-4502-4643-92b4-6fad825e76b4 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 760.921271] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3f2b299-c73b-4185-ac09-f9bced81b2a5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.927803] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b03284e-20d3-4c39-a1a4-f0553710b4a6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.933924] env[63531]: DEBUG nova.compute.manager [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] [instance: dccb2335-4220-4570-861b-16ff2d8f5a85] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 760.976418] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ab0dc3b-ccbf-4acc-91d3-b819dfe2f667 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.994390] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c274842-7c09-405e-97d4-da83db0d7625 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.998484] env[63531]: DEBUG oslo_vmware.api [None req-785bdba5-357d-4d51-af3f-e636dea49e89 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Task: {'id': task-1117676, 'name': PowerOffVM_Task, 'duration_secs': 0.243188} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.998739] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-785bdba5-357d-4d51-af3f-e636dea49e89 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 760.998909] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-785bdba5-357d-4d51-af3f-e636dea49e89 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 760.999787] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6bc33e21-8cc9-4144-b969-3499a784dbb7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.011086] env[63531]: DEBUG nova.compute.provider_tree [None req-db4fd4eb-d9b5-45f3-8b8c-2517c717ed2c tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 761.073378] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-785bdba5-357d-4d51-af3f-e636dea49e89 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 761.073902] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-785bdba5-357d-4d51-af3f-e636dea49e89 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 761.073902] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-785bdba5-357d-4d51-af3f-e636dea49e89 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Deleting the datastore file [datastore1] c3920e60-b27f-496b-b616-afb23ab0bc26 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 761.074280] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-79d264e4-fabf-4222-9948-3b5a128b7e2f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.077935] env[63531]: DEBUG nova.network.neutron [req-b11ca775-6519-40dc-8cc7-d400463b9865 req-b932a322-7c03-4cd1-815c-bfce8f79e0bb service nova] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Updated VIF entry in instance network info cache for port 91733d5a-84d6-48df-ba4d-7e362a642282. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 761.078369] env[63531]: DEBUG nova.network.neutron [req-b11ca775-6519-40dc-8cc7-d400463b9865 req-b932a322-7c03-4cd1-815c-bfce8f79e0bb service nova] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Updating instance_info_cache with network_info: [{"id": "91733d5a-84d6-48df-ba4d-7e362a642282", "address": "fa:16:3e:df:a1:31", "network": {"id": "d28b2087-329c-4917-aa6e-83ab83a63a74", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1054369059-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a634a61020b94d62a8fb37a3f96acc10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91733d5a-84", "ovs_interfaceid": "91733d5a-84d6-48df-ba4d-7e362a642282", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.080165] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 761.081055] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 761.088983] env[63531]: DEBUG oslo_vmware.api [None req-785bdba5-357d-4d51-af3f-e636dea49e89 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Waiting for the task: (returnval){ [ 761.088983] env[63531]: value = "task-1117678" [ 761.088983] env[63531]: _type = "Task" [ 761.088983] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.103612] env[63531]: DEBUG oslo_vmware.api [None req-785bdba5-357d-4d51-af3f-e636dea49e89 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Task: {'id': task-1117678, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.257163] env[63531]: DEBUG nova.network.neutron [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Successfully updated port: 262abfb7-7bcf-40e6-9ce3-b502de45b3b4 {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 761.261467] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Acquiring lock "43c29443-1210-4ee2-95c7-1257de308287" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 761.261772] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Lock "43c29443-1210-4ee2-95c7-1257de308287" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 761.516116] env[63531]: DEBUG nova.scheduler.client.report [None req-db4fd4eb-d9b5-45f3-8b8c-2517c717ed2c tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 761.565821] env[63531]: DEBUG nova.network.neutron [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Updating instance_info_cache with network_info: [{"id": "4c96ebbd-98fa-4d12-a388-b9d270b7aaae", "address": "fa:16:3e:97:f3:0b", "network": {"id": "1c410a92-1d7e-4b46-b7e0-7bcfd5dd7874", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1094381684", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.120", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bebde0e281d485ebdd70a1a3b613632", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d689fd7-f53e-4fd3-80d9-8d6b8fb7a164", "external-id": "nsx-vlan-transportzone-972", "segmentation_id": 972, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c96ebbd-98", "ovs_interfaceid": "4c96ebbd-98fa-4d12-a388-b9d270b7aaae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "5765492b-2d11-438b-b9a4-7514ce2ac73e", "address": "fa:16:3e:f7:f1:ba", "network": {"id": "485000a3-2bad-4c7d-b389-d4200550cc26", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-446289262", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.66", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "8bebde0e281d485ebdd70a1a3b613632", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5765492b-2d", "ovs_interfaceid": "5765492b-2d11-438b-b9a4-7514ce2ac73e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "4da738f4-b2e9-40c4-9a02-43b6ed658cfa", "address": "fa:16:3e:89:bd:7e", "network": {"id": "1c410a92-1d7e-4b46-b7e0-7bcfd5dd7874", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1094381684", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.105", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bebde0e281d485ebdd70a1a3b613632", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d689fd7-f53e-4fd3-80d9-8d6b8fb7a164", "external-id": "nsx-vlan-transportzone-972", "segmentation_id": 972, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4da738f4-b2", "ovs_interfaceid": "4da738f4-b2e9-40c4-9a02-43b6ed658cfa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.582575] env[63531]: DEBUG oslo_concurrency.lockutils [req-b11ca775-6519-40dc-8cc7-d400463b9865 req-b932a322-7c03-4cd1-815c-bfce8f79e0bb service nova] Releasing lock "refresh_cache-c3920e60-b27f-496b-b616-afb23ab0bc26" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 761.582760] env[63531]: DEBUG nova.compute.manager [req-b11ca775-6519-40dc-8cc7-d400463b9865 req-b932a322-7c03-4cd1-815c-bfce8f79e0bb service nova] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Received event network-vif-plugged-4da738f4-b2e9-40c4-9a02-43b6ed658cfa {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 761.583605] env[63531]: DEBUG oslo_concurrency.lockutils [req-b11ca775-6519-40dc-8cc7-d400463b9865 req-b932a322-7c03-4cd1-815c-bfce8f79e0bb service nova] Acquiring lock "e1693594-d8bb-4002-b71c-7013efb81d16-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 761.583605] env[63531]: DEBUG oslo_concurrency.lockutils [req-b11ca775-6519-40dc-8cc7-d400463b9865 req-b932a322-7c03-4cd1-815c-bfce8f79e0bb service nova] Lock "e1693594-d8bb-4002-b71c-7013efb81d16-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 761.583605] env[63531]: DEBUG oslo_concurrency.lockutils [req-b11ca775-6519-40dc-8cc7-d400463b9865 req-b932a322-7c03-4cd1-815c-bfce8f79e0bb service nova] Lock "e1693594-d8bb-4002-b71c-7013efb81d16-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 761.583605] env[63531]: DEBUG nova.compute.manager [req-b11ca775-6519-40dc-8cc7-d400463b9865 req-b932a322-7c03-4cd1-815c-bfce8f79e0bb service nova] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] No waiting events found dispatching network-vif-plugged-4da738f4-b2e9-40c4-9a02-43b6ed658cfa {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 761.583605] env[63531]: WARNING nova.compute.manager [req-b11ca775-6519-40dc-8cc7-d400463b9865 req-b932a322-7c03-4cd1-815c-bfce8f79e0bb service nova] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Received unexpected event network-vif-plugged-4da738f4-b2e9-40c4-9a02-43b6ed658cfa for instance with vm_state building and task_state spawning. [ 761.584328] env[63531]: DEBUG nova.compute.manager [req-b11ca775-6519-40dc-8cc7-d400463b9865 req-b932a322-7c03-4cd1-815c-bfce8f79e0bb service nova] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Received event network-changed-4da738f4-b2e9-40c4-9a02-43b6ed658cfa {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 761.584328] env[63531]: DEBUG nova.compute.manager [req-b11ca775-6519-40dc-8cc7-d400463b9865 req-b932a322-7c03-4cd1-815c-bfce8f79e0bb service nova] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Refreshing instance network info cache due to event network-changed-4da738f4-b2e9-40c4-9a02-43b6ed658cfa. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 761.584328] env[63531]: DEBUG oslo_concurrency.lockutils [req-b11ca775-6519-40dc-8cc7-d400463b9865 req-b932a322-7c03-4cd1-815c-bfce8f79e0bb service nova] Acquiring lock "refresh_cache-e1693594-d8bb-4002-b71c-7013efb81d16" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 761.591982] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0e74d968-ab93-4f14-92a6-24657e7c5585 tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Acquiring lock "e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 761.591982] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0e74d968-ab93-4f14-92a6-24657e7c5585 tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Lock "e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 761.591982] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0e74d968-ab93-4f14-92a6-24657e7c5585 tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Acquiring lock "e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 761.591982] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0e74d968-ab93-4f14-92a6-24657e7c5585 tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Lock "e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 761.591982] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0e74d968-ab93-4f14-92a6-24657e7c5585 tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Lock "e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 761.593917] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 761.593917] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Starting heal instance info cache {{(pid=63531) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 761.593917] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Rebuilding the list of instances to heal {{(pid=63531) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 761.600992] env[63531]: INFO nova.compute.manager [None req-0e74d968-ab93-4f14-92a6-24657e7c5585 tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74] Terminating instance [ 761.606958] env[63531]: DEBUG nova.compute.manager [None req-0e74d968-ab93-4f14-92a6-24657e7c5585 tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 761.607186] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-0e74d968-ab93-4f14-92a6-24657e7c5585 tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 761.607987] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e6e179d-f439-458f-af13-88f9f1b4d6e4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.618506] env[63531]: DEBUG oslo_vmware.api [None req-785bdba5-357d-4d51-af3f-e636dea49e89 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Task: {'id': task-1117678, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.232694} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.622457] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-785bdba5-357d-4d51-af3f-e636dea49e89 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 761.622457] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-785bdba5-357d-4d51-af3f-e636dea49e89 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 761.622457] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-785bdba5-357d-4d51-af3f-e636dea49e89 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 761.622457] env[63531]: INFO nova.compute.manager [None req-785bdba5-357d-4d51-af3f-e636dea49e89 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Took 1.18 seconds to destroy the instance on the hypervisor. [ 761.622457] env[63531]: DEBUG oslo.service.loopingcall [None req-785bdba5-357d-4d51-af3f-e636dea49e89 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 761.622689] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e74d968-ab93-4f14-92a6-24657e7c5585 tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 761.622770] env[63531]: DEBUG nova.compute.manager [-] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 761.622864] env[63531]: DEBUG nova.network.neutron [-] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 761.626131] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7b000a0a-7df3-42c1-ae27-30874a7899a4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.632429] env[63531]: DEBUG oslo_vmware.api [None req-0e74d968-ab93-4f14-92a6-24657e7c5585 tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Waiting for the task: (returnval){ [ 761.632429] env[63531]: value = "task-1117679" [ 761.632429] env[63531]: _type = "Task" [ 761.632429] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.641592] env[63531]: DEBUG oslo_vmware.api [None req-0e74d968-ab93-4f14-92a6-24657e7c5585 tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Task: {'id': task-1117679, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.762820] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "refresh_cache-29dd6de3-2765-408c-acc0-da47e5e0a977" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 761.763113] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquired lock "refresh_cache-29dd6de3-2765-408c-acc0-da47e5e0a977" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.763377] env[63531]: DEBUG nova.network.neutron [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 761.961508] env[63531]: DEBUG nova.compute.manager [req-312af46c-7e4d-47b9-9c2f-e0bf67d98480 req-b54a14bf-0741-404e-8ec3-15dbc6c6ea4e service nova] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Received event network-vif-deleted-91733d5a-84d6-48df-ba4d-7e362a642282 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 761.961725] env[63531]: INFO nova.compute.manager [req-312af46c-7e4d-47b9-9c2f-e0bf67d98480 req-b54a14bf-0741-404e-8ec3-15dbc6c6ea4e service nova] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Neutron deleted interface 91733d5a-84d6-48df-ba4d-7e362a642282; detaching it from the instance and deleting it from the info cache [ 761.961898] env[63531]: DEBUG nova.network.neutron [req-312af46c-7e4d-47b9-9c2f-e0bf67d98480 req-b54a14bf-0741-404e-8ec3-15dbc6c6ea4e service nova] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.982801] env[63531]: DEBUG nova.compute.manager [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] [instance: dccb2335-4220-4570-861b-16ff2d8f5a85] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 762.011327] env[63531]: DEBUG nova.virt.hardware [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 762.011591] env[63531]: DEBUG nova.virt.hardware [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 762.011747] env[63531]: DEBUG nova.virt.hardware [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 762.011932] env[63531]: DEBUG nova.virt.hardware [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 762.012129] env[63531]: DEBUG nova.virt.hardware [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 762.012302] env[63531]: DEBUG nova.virt.hardware [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 762.012505] env[63531]: DEBUG nova.virt.hardware [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 762.012664] env[63531]: DEBUG nova.virt.hardware [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 762.012829] env[63531]: DEBUG nova.virt.hardware [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 762.012988] env[63531]: DEBUG nova.virt.hardware [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 762.013178] env[63531]: DEBUG nova.virt.hardware [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 762.014225] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81d1eb47-c6b6-4f9e-8f8d-cd1c0f960217 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.023020] env[63531]: DEBUG oslo_concurrency.lockutils [None req-db4fd4eb-d9b5-45f3-8b8c-2517c717ed2c tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.105s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 762.023871] env[63531]: DEBUG oslo_concurrency.lockutils [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.635s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 762.025414] env[63531]: INFO nova.compute.claims [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 762.030614] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caeaed9f-0510-43bc-a808-689daf40ed22 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.056428] env[63531]: INFO nova.scheduler.client.report [None req-db4fd4eb-d9b5-45f3-8b8c-2517c717ed2c tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Deleted allocations for instance c3d597b4-1dc3-48d8-9bee-e73c8929181b [ 762.072201] env[63531]: DEBUG nova.compute.manager [req-9d9031a8-8d68-4380-b818-ef232154f65f req-ed1efd3b-d60d-4029-b0f1-ca33cdfac901 service nova] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Received event network-vif-plugged-262abfb7-7bcf-40e6-9ce3-b502de45b3b4 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 762.072422] env[63531]: DEBUG oslo_concurrency.lockutils [req-9d9031a8-8d68-4380-b818-ef232154f65f req-ed1efd3b-d60d-4029-b0f1-ca33cdfac901 service nova] Acquiring lock "29dd6de3-2765-408c-acc0-da47e5e0a977-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 762.072639] env[63531]: DEBUG oslo_concurrency.lockutils [req-9d9031a8-8d68-4380-b818-ef232154f65f req-ed1efd3b-d60d-4029-b0f1-ca33cdfac901 service nova] Lock "29dd6de3-2765-408c-acc0-da47e5e0a977-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 762.072800] env[63531]: DEBUG oslo_concurrency.lockutils [req-9d9031a8-8d68-4380-b818-ef232154f65f req-ed1efd3b-d60d-4029-b0f1-ca33cdfac901 service nova] Lock "29dd6de3-2765-408c-acc0-da47e5e0a977-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 762.072963] env[63531]: DEBUG nova.compute.manager [req-9d9031a8-8d68-4380-b818-ef232154f65f req-ed1efd3b-d60d-4029-b0f1-ca33cdfac901 service nova] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] No waiting events found dispatching network-vif-plugged-262abfb7-7bcf-40e6-9ce3-b502de45b3b4 {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 762.073175] env[63531]: WARNING nova.compute.manager [req-9d9031a8-8d68-4380-b818-ef232154f65f req-ed1efd3b-d60d-4029-b0f1-ca33cdfac901 service nova] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Received unexpected event network-vif-plugged-262abfb7-7bcf-40e6-9ce3-b502de45b3b4 for instance with vm_state building and task_state spawning. [ 762.073295] env[63531]: DEBUG nova.compute.manager [req-9d9031a8-8d68-4380-b818-ef232154f65f req-ed1efd3b-d60d-4029-b0f1-ca33cdfac901 service nova] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Received event network-changed-262abfb7-7bcf-40e6-9ce3-b502de45b3b4 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 762.073442] env[63531]: DEBUG nova.compute.manager [req-9d9031a8-8d68-4380-b818-ef232154f65f req-ed1efd3b-d60d-4029-b0f1-ca33cdfac901 service nova] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Refreshing instance network info cache due to event network-changed-262abfb7-7bcf-40e6-9ce3-b502de45b3b4. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 762.073603] env[63531]: DEBUG oslo_concurrency.lockutils [req-9d9031a8-8d68-4380-b818-ef232154f65f req-ed1efd3b-d60d-4029-b0f1-ca33cdfac901 service nova] Acquiring lock "refresh_cache-29dd6de3-2765-408c-acc0-da47e5e0a977" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 762.073959] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Releasing lock "refresh_cache-e1693594-d8bb-4002-b71c-7013efb81d16" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 762.074423] env[63531]: DEBUG nova.compute.manager [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Instance network_info: |[{"id": "4c96ebbd-98fa-4d12-a388-b9d270b7aaae", "address": "fa:16:3e:97:f3:0b", "network": {"id": "1c410a92-1d7e-4b46-b7e0-7bcfd5dd7874", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1094381684", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.120", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bebde0e281d485ebdd70a1a3b613632", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d689fd7-f53e-4fd3-80d9-8d6b8fb7a164", "external-id": "nsx-vlan-transportzone-972", "segmentation_id": 972, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c96ebbd-98", "ovs_interfaceid": "4c96ebbd-98fa-4d12-a388-b9d270b7aaae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "5765492b-2d11-438b-b9a4-7514ce2ac73e", "address": "fa:16:3e:f7:f1:ba", "network": {"id": "485000a3-2bad-4c7d-b389-d4200550cc26", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-446289262", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.66", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "8bebde0e281d485ebdd70a1a3b613632", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5765492b-2d", "ovs_interfaceid": "5765492b-2d11-438b-b9a4-7514ce2ac73e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "4da738f4-b2e9-40c4-9a02-43b6ed658cfa", "address": "fa:16:3e:89:bd:7e", "network": {"id": "1c410a92-1d7e-4b46-b7e0-7bcfd5dd7874", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1094381684", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.105", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bebde0e281d485ebdd70a1a3b613632", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d689fd7-f53e-4fd3-80d9-8d6b8fb7a164", "external-id": "nsx-vlan-transportzone-972", "segmentation_id": 972, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4da738f4-b2", "ovs_interfaceid": "4da738f4-b2e9-40c4-9a02-43b6ed658cfa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 762.075501] env[63531]: DEBUG oslo_concurrency.lockutils [req-b11ca775-6519-40dc-8cc7-d400463b9865 req-b932a322-7c03-4cd1-815c-bfce8f79e0bb service nova] Acquired lock "refresh_cache-e1693594-d8bb-4002-b71c-7013efb81d16" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.075501] env[63531]: DEBUG nova.network.neutron [req-b11ca775-6519-40dc-8cc7-d400463b9865 req-b932a322-7c03-4cd1-815c-bfce8f79e0bb service nova] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Refreshing network info cache for port 4da738f4-b2e9-40c4-9a02-43b6ed658cfa {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 762.076219] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:97:f3:0b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7d689fd7-f53e-4fd3-80d9-8d6b8fb7a164', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4c96ebbd-98fa-4d12-a388-b9d270b7aaae', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:f7:f1:ba', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fb224918-e294-4b76-80f9-2fa0031b7dc2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5765492b-2d11-438b-b9a4-7514ce2ac73e', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:89:bd:7e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7d689fd7-f53e-4fd3-80d9-8d6b8fb7a164', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4da738f4-b2e9-40c4-9a02-43b6ed658cfa', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 762.089740] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Creating folder: Project (8bebde0e281d485ebdd70a1a3b613632). Parent ref: group-v244585. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 762.089740] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d5332129-e51f-4859-be62-3eb4d8654144 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.102682] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Skipping network cache update for instance because it is being deleted. {{(pid=63531) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 762.102854] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74] Skipping network cache update for instance because it is being deleted. {{(pid=63531) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 762.102987] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Skipping network cache update for instance because it is Building. {{(pid=63531) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 762.103124] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Skipping network cache update for instance because it is Building. {{(pid=63531) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 762.103248] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: dccb2335-4220-4570-861b-16ff2d8f5a85] Skipping network cache update for instance because it is Building. {{(pid=63531) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 762.105369] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Created folder: Project (8bebde0e281d485ebdd70a1a3b613632) in parent group-v244585. [ 762.105548] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Creating folder: Instances. Parent ref: group-v244640. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 762.106056] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d8fa6b69-3f4c-4e7e-8320-25d5b95f1f88 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.117323] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Created folder: Instances in parent group-v244640. [ 762.117559] env[63531]: DEBUG oslo.service.loopingcall [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 762.117741] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 762.117936] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f75a2c4b-088c-4e1e-b5a4-1f97e6b204d1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.137163] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Acquiring lock "refresh_cache-c3d597b4-1dc3-48d8-9bee-e73c8929181b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 762.137308] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Acquired lock "refresh_cache-c3d597b4-1dc3-48d8-9bee-e73c8929181b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.137450] env[63531]: DEBUG nova.network.neutron [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Forcefully refreshing network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 762.137603] env[63531]: DEBUG nova.objects.instance [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Lazy-loading 'info_cache' on Instance uuid c3d597b4-1dc3-48d8-9bee-e73c8929181b {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 762.147468] env[63531]: DEBUG oslo_vmware.api [None req-0e74d968-ab93-4f14-92a6-24657e7c5585 tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Task: {'id': task-1117679, 'name': PowerOffVM_Task, 'duration_secs': 0.205344} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.148642] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e74d968-ab93-4f14-92a6-24657e7c5585 tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 762.148806] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-0e74d968-ab93-4f14-92a6-24657e7c5585 tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 762.149027] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 762.149027] env[63531]: value = "task-1117682" [ 762.149027] env[63531]: _type = "Task" [ 762.149027] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.149213] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-55842f3b-6922-4992-bee8-f37971fa3770 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.160584] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117682, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.223122] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-0e74d968-ab93-4f14-92a6-24657e7c5585 tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 762.223525] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-0e74d968-ab93-4f14-92a6-24657e7c5585 tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 762.223630] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e74d968-ab93-4f14-92a6-24657e7c5585 tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Deleting the datastore file [datastore1] e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 762.223872] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-68fdf11c-2682-4645-9c07-0a016cf6aeca {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.230892] env[63531]: DEBUG oslo_vmware.api [None req-0e74d968-ab93-4f14-92a6-24657e7c5585 tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Waiting for the task: (returnval){ [ 762.230892] env[63531]: value = "task-1117684" [ 762.230892] env[63531]: _type = "Task" [ 762.230892] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.239430] env[63531]: DEBUG oslo_vmware.api [None req-0e74d968-ab93-4f14-92a6-24657e7c5585 tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Task: {'id': task-1117684, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.309601] env[63531]: DEBUG nova.network.neutron [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 762.386955] env[63531]: DEBUG nova.network.neutron [-] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.464554] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-56b6d37c-97de-4bfe-8e06-80018fb250ff {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.475060] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0055cbc2-ffe2-4dbe-96c5-bf895319b8c3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.507817] env[63531]: DEBUG nova.compute.manager [req-312af46c-7e4d-47b9-9c2f-e0bf67d98480 req-b54a14bf-0741-404e-8ec3-15dbc6c6ea4e service nova] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Detach interface failed, port_id=91733d5a-84d6-48df-ba4d-7e362a642282, reason: Instance c3920e60-b27f-496b-b616-afb23ab0bc26 could not be found. {{(pid=63531) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 762.566487] env[63531]: DEBUG oslo_concurrency.lockutils [None req-db4fd4eb-d9b5-45f3-8b8c-2517c717ed2c tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Lock "c3d597b4-1dc3-48d8-9bee-e73c8929181b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.849s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 762.583528] env[63531]: DEBUG nova.network.neutron [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Updating instance_info_cache with network_info: [{"id": "262abfb7-7bcf-40e6-9ce3-b502de45b3b4", "address": "fa:16:3e:0e:00:c9", "network": {"id": "0667c176-cfdf-4e43-b152-e00bdc224448", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-243715985-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "308eabafea9c4188a58a0f1c22074d2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8f580e6-1d86-41ee-9ebe-c531cb9299c6", "external-id": "nsx-vlan-transportzone-150", "segmentation_id": 150, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap262abfb7-7b", "ovs_interfaceid": "262abfb7-7bcf-40e6-9ce3-b502de45b3b4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.665976] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117682, 'name': CreateVM_Task, 'duration_secs': 0.492365} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.665976] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 762.665976] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 762.665976] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.665976] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 762.670447] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4d6b5023-22f1-4b88-9856-01697882f549 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.675955] env[63531]: DEBUG oslo_vmware.api [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Waiting for the task: (returnval){ [ 762.675955] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52156f27-d83d-5c36-46f4-c5a4f3b4fdf5" [ 762.675955] env[63531]: _type = "Task" [ 762.675955] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.684410] env[63531]: DEBUG oslo_vmware.api [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52156f27-d83d-5c36-46f4-c5a4f3b4fdf5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.746466] env[63531]: DEBUG oslo_vmware.api [None req-0e74d968-ab93-4f14-92a6-24657e7c5585 tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Task: {'id': task-1117684, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.267223} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.746466] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e74d968-ab93-4f14-92a6-24657e7c5585 tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 762.746466] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-0e74d968-ab93-4f14-92a6-24657e7c5585 tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 762.746466] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-0e74d968-ab93-4f14-92a6-24657e7c5585 tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 762.746466] env[63531]: INFO nova.compute.manager [None req-0e74d968-ab93-4f14-92a6-24657e7c5585 tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] [instance: e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74] Took 1.14 seconds to destroy the instance on the hypervisor. [ 762.746466] env[63531]: DEBUG oslo.service.loopingcall [None req-0e74d968-ab93-4f14-92a6-24657e7c5585 tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 762.746466] env[63531]: DEBUG nova.compute.manager [-] [instance: e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 762.746466] env[63531]: DEBUG nova.network.neutron [-] [instance: e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 762.894414] env[63531]: INFO nova.compute.manager [-] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Took 1.27 seconds to deallocate network for instance. [ 762.951175] env[63531]: DEBUG nova.network.neutron [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] [instance: dccb2335-4220-4570-861b-16ff2d8f5a85] Successfully updated port: d69ccbf2-4502-4643-92b4-6fad825e76b4 {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 762.996356] env[63531]: DEBUG nova.network.neutron [req-b11ca775-6519-40dc-8cc7-d400463b9865 req-b932a322-7c03-4cd1-815c-bfce8f79e0bb service nova] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Updated VIF entry in instance network info cache for port 4da738f4-b2e9-40c4-9a02-43b6ed658cfa. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 762.996356] env[63531]: DEBUG nova.network.neutron [req-b11ca775-6519-40dc-8cc7-d400463b9865 req-b932a322-7c03-4cd1-815c-bfce8f79e0bb service nova] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Updating instance_info_cache with network_info: [{"id": "4c96ebbd-98fa-4d12-a388-b9d270b7aaae", "address": "fa:16:3e:97:f3:0b", "network": {"id": "1c410a92-1d7e-4b46-b7e0-7bcfd5dd7874", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1094381684", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.120", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bebde0e281d485ebdd70a1a3b613632", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d689fd7-f53e-4fd3-80d9-8d6b8fb7a164", "external-id": "nsx-vlan-transportzone-972", "segmentation_id": 972, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c96ebbd-98", "ovs_interfaceid": "4c96ebbd-98fa-4d12-a388-b9d270b7aaae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "5765492b-2d11-438b-b9a4-7514ce2ac73e", "address": "fa:16:3e:f7:f1:ba", "network": {"id": "485000a3-2bad-4c7d-b389-d4200550cc26", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-446289262", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.66", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "8bebde0e281d485ebdd70a1a3b613632", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5765492b-2d", "ovs_interfaceid": "5765492b-2d11-438b-b9a4-7514ce2ac73e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "4da738f4-b2e9-40c4-9a02-43b6ed658cfa", "address": "fa:16:3e:89:bd:7e", "network": {"id": "1c410a92-1d7e-4b46-b7e0-7bcfd5dd7874", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1094381684", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.105", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bebde0e281d485ebdd70a1a3b613632", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d689fd7-f53e-4fd3-80d9-8d6b8fb7a164", "external-id": "nsx-vlan-transportzone-972", "segmentation_id": 972, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4da738f4-b2", "ovs_interfaceid": "4da738f4-b2e9-40c4-9a02-43b6ed658cfa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 763.088530] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Releasing lock "refresh_cache-29dd6de3-2765-408c-acc0-da47e5e0a977" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 763.088893] env[63531]: DEBUG nova.compute.manager [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Instance network_info: |[{"id": "262abfb7-7bcf-40e6-9ce3-b502de45b3b4", "address": "fa:16:3e:0e:00:c9", "network": {"id": "0667c176-cfdf-4e43-b152-e00bdc224448", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-243715985-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "308eabafea9c4188a58a0f1c22074d2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8f580e6-1d86-41ee-9ebe-c531cb9299c6", "external-id": "nsx-vlan-transportzone-150", "segmentation_id": 150, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap262abfb7-7b", "ovs_interfaceid": "262abfb7-7bcf-40e6-9ce3-b502de45b3b4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 763.089215] env[63531]: DEBUG oslo_concurrency.lockutils [req-9d9031a8-8d68-4380-b818-ef232154f65f req-ed1efd3b-d60d-4029-b0f1-ca33cdfac901 service nova] Acquired lock "refresh_cache-29dd6de3-2765-408c-acc0-da47e5e0a977" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.089392] env[63531]: DEBUG nova.network.neutron [req-9d9031a8-8d68-4380-b818-ef232154f65f req-ed1efd3b-d60d-4029-b0f1-ca33cdfac901 service nova] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Refreshing network info cache for port 262abfb7-7bcf-40e6-9ce3-b502de45b3b4 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 763.090668] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0e:00:c9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c8f580e6-1d86-41ee-9ebe-c531cb9299c6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '262abfb7-7bcf-40e6-9ce3-b502de45b3b4', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 763.104010] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Creating folder: Project (308eabafea9c4188a58a0f1c22074d2f). Parent ref: group-v244585. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 763.108011] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1df1808d-cc74-4633-b181-cd7d1fd215fc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.120536] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Created folder: Project (308eabafea9c4188a58a0f1c22074d2f) in parent group-v244585. [ 763.120536] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Creating folder: Instances. Parent ref: group-v244643. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 763.120536] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3f890325-db77-4d58-9714-b1c22bcf9b2b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.132846] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Created folder: Instances in parent group-v244643. [ 763.133942] env[63531]: DEBUG oslo.service.loopingcall [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 763.134172] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 763.134381] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-297e44c5-7dc6-4e44-a01a-a68d1285259a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.161934] env[63531]: DEBUG nova.compute.utils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Can not refresh info_cache because instance was not found {{(pid=63531) refresh_info_cache_for_instance /opt/stack/nova/nova/compute/utils.py:1024}} [ 763.172529] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 763.172529] env[63531]: value = "task-1117687" [ 763.172529] env[63531]: _type = "Task" [ 763.172529] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.190075] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117687, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.190984] env[63531]: DEBUG nova.network.neutron [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 763.206233] env[63531]: DEBUG oslo_vmware.api [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52156f27-d83d-5c36-46f4-c5a4f3b4fdf5, 'name': SearchDatastore_Task, 'duration_secs': 0.015342} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.206313] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 763.206897] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 763.206897] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 763.206897] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.207941] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 763.207941] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-93dfa8d4-42e5-45f7-9e85-bd1401b3a398 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.222756] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 763.222961] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 763.224386] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b49e8ced-871d-4707-ae07-7f7abf3e3342 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.232509] env[63531]: DEBUG oslo_vmware.api [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Waiting for the task: (returnval){ [ 763.232509] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5291ac6d-976c-3c41-9ed4-59878561a2bf" [ 763.232509] env[63531]: _type = "Task" [ 763.232509] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.241416] env[63531]: DEBUG oslo_vmware.api [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5291ac6d-976c-3c41-9ed4-59878561a2bf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.402183] env[63531]: DEBUG oslo_concurrency.lockutils [None req-785bdba5-357d-4d51-af3f-e636dea49e89 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 763.460069] env[63531]: DEBUG oslo_concurrency.lockutils [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Acquiring lock "refresh_cache-dccb2335-4220-4570-861b-16ff2d8f5a85" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 763.460069] env[63531]: DEBUG oslo_concurrency.lockutils [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Acquired lock "refresh_cache-dccb2335-4220-4570-861b-16ff2d8f5a85" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.460069] env[63531]: DEBUG nova.network.neutron [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] [instance: dccb2335-4220-4570-861b-16ff2d8f5a85] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 763.498774] env[63531]: DEBUG oslo_concurrency.lockutils [req-b11ca775-6519-40dc-8cc7-d400463b9865 req-b932a322-7c03-4cd1-815c-bfce8f79e0bb service nova] Releasing lock "refresh_cache-e1693594-d8bb-4002-b71c-7013efb81d16" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 763.529083] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-275cd002-d145-4cf1-8c0b-2bbac498bb42 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.536169] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e17d9cb0-b010-4dc5-b066-417d555a20d1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.569196] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bc2aaf8-8540-4951-b984-4f185918cfd3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.577586] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-522644f9-2eed-40a5-8554-0aba54a5bd64 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.591328] env[63531]: DEBUG nova.compute.provider_tree [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Updating inventory in ProviderTree for provider 9996a314-a62e-4282-8252-f387e7e3fabb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 763.669248] env[63531]: DEBUG nova.network.neutron [-] [instance: e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 763.685124] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117687, 'name': CreateVM_Task, 'duration_secs': 0.388279} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.685256] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 763.685936] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 763.686114] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.686425] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 763.686675] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4d6e7e44-973e-42f0-9835-4b960f9a72b6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.691679] env[63531]: DEBUG oslo_vmware.api [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 763.691679] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52557120-c341-e8de-471e-ab9cf6ee9deb" [ 763.691679] env[63531]: _type = "Task" [ 763.691679] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.704022] env[63531]: DEBUG oslo_vmware.api [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52557120-c341-e8de-471e-ab9cf6ee9deb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.748231] env[63531]: DEBUG oslo_vmware.api [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5291ac6d-976c-3c41-9ed4-59878561a2bf, 'name': SearchDatastore_Task, 'duration_secs': 0.010442} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.749064] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eae17fdf-fd6c-4bf3-a724-259f9aaaa91b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.754325] env[63531]: DEBUG oslo_vmware.api [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Waiting for the task: (returnval){ [ 763.754325] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52913d77-1cdf-a6b0-c812-e978d7c0df9f" [ 763.754325] env[63531]: _type = "Task" [ 763.754325] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.762337] env[63531]: DEBUG oslo_vmware.api [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52913d77-1cdf-a6b0-c812-e978d7c0df9f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.843824] env[63531]: DEBUG nova.network.neutron [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.005276] env[63531]: DEBUG nova.network.neutron [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] [instance: dccb2335-4220-4570-861b-16ff2d8f5a85] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 764.014724] env[63531]: DEBUG nova.network.neutron [req-9d9031a8-8d68-4380-b818-ef232154f65f req-ed1efd3b-d60d-4029-b0f1-ca33cdfac901 service nova] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Updated VIF entry in instance network info cache for port 262abfb7-7bcf-40e6-9ce3-b502de45b3b4. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 764.015092] env[63531]: DEBUG nova.network.neutron [req-9d9031a8-8d68-4380-b818-ef232154f65f req-ed1efd3b-d60d-4029-b0f1-ca33cdfac901 service nova] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Updating instance_info_cache with network_info: [{"id": "262abfb7-7bcf-40e6-9ce3-b502de45b3b4", "address": "fa:16:3e:0e:00:c9", "network": {"id": "0667c176-cfdf-4e43-b152-e00bdc224448", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-243715985-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "308eabafea9c4188a58a0f1c22074d2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8f580e6-1d86-41ee-9ebe-c531cb9299c6", "external-id": "nsx-vlan-transportzone-150", "segmentation_id": 150, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap262abfb7-7b", "ovs_interfaceid": "262abfb7-7bcf-40e6-9ce3-b502de45b3b4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.052502] env[63531]: DEBUG nova.compute.manager [req-85da9566-24fa-4902-8ebb-de25e2e47a83 req-4e521678-f7cd-447d-bd32-97aa28a15146 service nova] [instance: dccb2335-4220-4570-861b-16ff2d8f5a85] Received event network-vif-plugged-d69ccbf2-4502-4643-92b4-6fad825e76b4 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 764.052716] env[63531]: DEBUG oslo_concurrency.lockutils [req-85da9566-24fa-4902-8ebb-de25e2e47a83 req-4e521678-f7cd-447d-bd32-97aa28a15146 service nova] Acquiring lock "dccb2335-4220-4570-861b-16ff2d8f5a85-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 764.052914] env[63531]: DEBUG oslo_concurrency.lockutils [req-85da9566-24fa-4902-8ebb-de25e2e47a83 req-4e521678-f7cd-447d-bd32-97aa28a15146 service nova] Lock "dccb2335-4220-4570-861b-16ff2d8f5a85-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 764.053156] env[63531]: DEBUG oslo_concurrency.lockutils [req-85da9566-24fa-4902-8ebb-de25e2e47a83 req-4e521678-f7cd-447d-bd32-97aa28a15146 service nova] Lock "dccb2335-4220-4570-861b-16ff2d8f5a85-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 764.053300] env[63531]: DEBUG nova.compute.manager [req-85da9566-24fa-4902-8ebb-de25e2e47a83 req-4e521678-f7cd-447d-bd32-97aa28a15146 service nova] [instance: dccb2335-4220-4570-861b-16ff2d8f5a85] No waiting events found dispatching network-vif-plugged-d69ccbf2-4502-4643-92b4-6fad825e76b4 {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 764.053406] env[63531]: WARNING nova.compute.manager [req-85da9566-24fa-4902-8ebb-de25e2e47a83 req-4e521678-f7cd-447d-bd32-97aa28a15146 service nova] [instance: dccb2335-4220-4570-861b-16ff2d8f5a85] Received unexpected event network-vif-plugged-d69ccbf2-4502-4643-92b4-6fad825e76b4 for instance with vm_state building and task_state spawning. [ 764.053589] env[63531]: DEBUG nova.compute.manager [req-85da9566-24fa-4902-8ebb-de25e2e47a83 req-4e521678-f7cd-447d-bd32-97aa28a15146 service nova] [instance: dccb2335-4220-4570-861b-16ff2d8f5a85] Received event network-changed-d69ccbf2-4502-4643-92b4-6fad825e76b4 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 764.053692] env[63531]: DEBUG nova.compute.manager [req-85da9566-24fa-4902-8ebb-de25e2e47a83 req-4e521678-f7cd-447d-bd32-97aa28a15146 service nova] [instance: dccb2335-4220-4570-861b-16ff2d8f5a85] Refreshing instance network info cache due to event network-changed-d69ccbf2-4502-4643-92b4-6fad825e76b4. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 764.053836] env[63531]: DEBUG oslo_concurrency.lockutils [req-85da9566-24fa-4902-8ebb-de25e2e47a83 req-4e521678-f7cd-447d-bd32-97aa28a15146 service nova] Acquiring lock "refresh_cache-dccb2335-4220-4570-861b-16ff2d8f5a85" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 764.112584] env[63531]: ERROR nova.scheduler.client.report [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [req-64d265cf-0ca1-4b94-af41-31ca5f3fb904] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 9996a314-a62e-4282-8252-f387e7e3fabb. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-64d265cf-0ca1-4b94-af41-31ca5f3fb904"}]} [ 764.129190] env[63531]: DEBUG nova.scheduler.client.report [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Refreshing inventories for resource provider 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 764.145453] env[63531]: DEBUG nova.scheduler.client.report [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Updating ProviderTree inventory for provider 9996a314-a62e-4282-8252-f387e7e3fabb from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 764.145715] env[63531]: DEBUG nova.compute.provider_tree [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Updating inventory in ProviderTree for provider 9996a314-a62e-4282-8252-f387e7e3fabb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 764.164841] env[63531]: DEBUG nova.scheduler.client.report [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Refreshing aggregate associations for resource provider 9996a314-a62e-4282-8252-f387e7e3fabb, aggregates: None {{(pid=63531) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 764.172205] env[63531]: INFO nova.compute.manager [-] [instance: e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74] Took 1.43 seconds to deallocate network for instance. [ 764.186923] env[63531]: DEBUG nova.compute.manager [req-cfc88429-de4f-4d7f-991b-dbddd05bd6bb req-fd95e6ff-1bcd-4244-a57e-d31a7d25f759 service nova] [instance: e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74] Received event network-vif-deleted-8a0490e2-e2d3-4316-9682-7d9a24122671 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 764.197474] env[63531]: DEBUG nova.scheduler.client.report [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Refreshing trait associations for resource provider 9996a314-a62e-4282-8252-f387e7e3fabb, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE {{(pid=63531) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 764.203287] env[63531]: DEBUG oslo_vmware.api [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52557120-c341-e8de-471e-ab9cf6ee9deb, 'name': SearchDatastore_Task, 'duration_secs': 0.031704} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.203573] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 764.203802] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 764.204048] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 764.204441] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 764.204441] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 764.204910] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3f7d3dcd-6807-4de1-93ec-836be9ecb009 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.215674] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 764.215901] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 764.216598] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-364c6618-6e41-4dac-ad10-aa03f48f3dab {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.227250] env[63531]: DEBUG oslo_vmware.api [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 764.227250] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52c01fed-b858-df6c-7812-f48463d2ede6" [ 764.227250] env[63531]: _type = "Task" [ 764.227250] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.238440] env[63531]: DEBUG oslo_vmware.api [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52c01fed-b858-df6c-7812-f48463d2ede6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.269713] env[63531]: DEBUG oslo_vmware.api [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52913d77-1cdf-a6b0-c812-e978d7c0df9f, 'name': SearchDatastore_Task, 'duration_secs': 0.025151} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.273444] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 764.273979] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] e1693594-d8bb-4002-b71c-7013efb81d16/e1693594-d8bb-4002-b71c-7013efb81d16.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 764.274599] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-05382ba2-c9ad-4e22-9d9a-9c010af1c168 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.281643] env[63531]: DEBUG oslo_vmware.api [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Waiting for the task: (returnval){ [ 764.281643] env[63531]: value = "task-1117688" [ 764.281643] env[63531]: _type = "Task" [ 764.281643] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.282749] env[63531]: DEBUG nova.network.neutron [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] [instance: dccb2335-4220-4570-861b-16ff2d8f5a85] Updating instance_info_cache with network_info: [{"id": "d69ccbf2-4502-4643-92b4-6fad825e76b4", "address": "fa:16:3e:43:06:ca", "network": {"id": "559568d9-859c-4dcf-9a3b-167f5dd3055c", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-251572816-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c4d75d29af5145bfb5140aabaeabf7e0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f880ac2e-d532-4f54-87bb-998a8d1bca78", "external-id": "nsx-vlan-transportzone-491", "segmentation_id": 491, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd69ccbf2-45", "ovs_interfaceid": "d69ccbf2-4502-4643-92b4-6fad825e76b4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.296244] env[63531]: DEBUG oslo_vmware.api [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Task: {'id': task-1117688, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.349076] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Releasing lock "refresh_cache-c3d597b4-1dc3-48d8-9bee-e73c8929181b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 764.349281] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Updated the network info_cache for instance {{(pid=63531) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 764.350299] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 764.350547] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 764.350769] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 764.351012] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 764.351202] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 764.351407] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 764.351615] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63531) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 764.351835] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager.update_available_resource {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 764.519118] env[63531]: DEBUG oslo_concurrency.lockutils [req-9d9031a8-8d68-4380-b818-ef232154f65f req-ed1efd3b-d60d-4029-b0f1-ca33cdfac901 service nova] Releasing lock "refresh_cache-29dd6de3-2765-408c-acc0-da47e5e0a977" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 764.586943] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb618880-e1e1-450d-b766-93cde95ac7c7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.596028] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da78350d-b9cb-4d96-9f9b-97a179331276 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.638525] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f4d2120-4264-4d73-bf3b-7acb01ab172c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.647725] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5638c6c5-ea22-4e6d-bc68-1e8c2d90353e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.665320] env[63531]: DEBUG nova.compute.provider_tree [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Updating inventory in ProviderTree for provider 9996a314-a62e-4282-8252-f387e7e3fabb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 764.681253] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0e74d968-ab93-4f14-92a6-24657e7c5585 tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 764.740596] env[63531]: DEBUG oslo_vmware.api [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52c01fed-b858-df6c-7812-f48463d2ede6, 'name': SearchDatastore_Task, 'duration_secs': 0.033525} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.741176] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d2830ae9-7a9b-49b2-b432-9710226f3c86 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.747559] env[63531]: DEBUG oslo_vmware.api [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 764.747559] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52d96786-a77c-c1ae-1129-5e0f6ff6c104" [ 764.747559] env[63531]: _type = "Task" [ 764.747559] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.756077] env[63531]: DEBUG oslo_vmware.api [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52d96786-a77c-c1ae-1129-5e0f6ff6c104, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.792095] env[63531]: DEBUG oslo_concurrency.lockutils [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Releasing lock "refresh_cache-dccb2335-4220-4570-861b-16ff2d8f5a85" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 764.792356] env[63531]: DEBUG nova.compute.manager [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] [instance: dccb2335-4220-4570-861b-16ff2d8f5a85] Instance network_info: |[{"id": "d69ccbf2-4502-4643-92b4-6fad825e76b4", "address": "fa:16:3e:43:06:ca", "network": {"id": "559568d9-859c-4dcf-9a3b-167f5dd3055c", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-251572816-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c4d75d29af5145bfb5140aabaeabf7e0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f880ac2e-d532-4f54-87bb-998a8d1bca78", "external-id": "nsx-vlan-transportzone-491", "segmentation_id": 491, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd69ccbf2-45", "ovs_interfaceid": "d69ccbf2-4502-4643-92b4-6fad825e76b4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 764.795852] env[63531]: DEBUG oslo_concurrency.lockutils [req-85da9566-24fa-4902-8ebb-de25e2e47a83 req-4e521678-f7cd-447d-bd32-97aa28a15146 service nova] Acquired lock "refresh_cache-dccb2335-4220-4570-861b-16ff2d8f5a85" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 764.796113] env[63531]: DEBUG nova.network.neutron [req-85da9566-24fa-4902-8ebb-de25e2e47a83 req-4e521678-f7cd-447d-bd32-97aa28a15146 service nova] [instance: dccb2335-4220-4570-861b-16ff2d8f5a85] Refreshing network info cache for port d69ccbf2-4502-4643-92b4-6fad825e76b4 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 764.797455] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] [instance: dccb2335-4220-4570-861b-16ff2d8f5a85] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:43:06:ca', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f880ac2e-d532-4f54-87bb-998a8d1bca78', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd69ccbf2-4502-4643-92b4-6fad825e76b4', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 764.806531] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Creating folder: Project (c4d75d29af5145bfb5140aabaeabf7e0). Parent ref: group-v244585. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 764.806922] env[63531]: DEBUG oslo_vmware.api [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Task: {'id': task-1117688, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.810079] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b23aec73-9486-4b7a-ae67-1f8c4de2c9da {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.820831] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Created folder: Project (c4d75d29af5145bfb5140aabaeabf7e0) in parent group-v244585. [ 764.821046] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Creating folder: Instances. Parent ref: group-v244646. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 764.821288] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-63091bde-d964-4fcf-99c2-c2dbd57d32c7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.832163] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Created folder: Instances in parent group-v244646. [ 764.832163] env[63531]: DEBUG oslo.service.loopingcall [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 764.832163] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dccb2335-4220-4570-861b-16ff2d8f5a85] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 764.832163] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-23813279-02a5-4549-bedb-5a0e5a1edf49 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.849264] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 764.849264] env[63531]: value = "task-1117691" [ 764.849264] env[63531]: _type = "Task" [ 764.849264] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.854420] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 764.857374] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117691, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.104730] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c52c5465-cb3f-44ae-a7bc-34ca246641a2 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Acquiring lock "645abb7a-cf73-4fb0-a9a0-49205f060d6f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 765.105033] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c52c5465-cb3f-44ae-a7bc-34ca246641a2 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Lock "645abb7a-cf73-4fb0-a9a0-49205f060d6f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 765.105278] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c52c5465-cb3f-44ae-a7bc-34ca246641a2 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Acquiring lock "645abb7a-cf73-4fb0-a9a0-49205f060d6f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 765.105491] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c52c5465-cb3f-44ae-a7bc-34ca246641a2 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Lock "645abb7a-cf73-4fb0-a9a0-49205f060d6f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 765.105697] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c52c5465-cb3f-44ae-a7bc-34ca246641a2 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Lock "645abb7a-cf73-4fb0-a9a0-49205f060d6f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 765.108237] env[63531]: INFO nova.compute.manager [None req-c52c5465-cb3f-44ae-a7bc-34ca246641a2 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: 645abb7a-cf73-4fb0-a9a0-49205f060d6f] Terminating instance [ 765.115021] env[63531]: DEBUG nova.compute.manager [None req-c52c5465-cb3f-44ae-a7bc-34ca246641a2 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: 645abb7a-cf73-4fb0-a9a0-49205f060d6f] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 765.115021] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c52c5465-cb3f-44ae-a7bc-34ca246641a2 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: 645abb7a-cf73-4fb0-a9a0-49205f060d6f] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 765.115021] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-022376b5-8274-4d61-adac-029a81e4b2e0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.127235] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-c52c5465-cb3f-44ae-a7bc-34ca246641a2 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: 645abb7a-cf73-4fb0-a9a0-49205f060d6f] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 765.127471] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-58f04640-d9ee-4e99-910d-ed395406667d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.134207] env[63531]: DEBUG oslo_vmware.api [None req-c52c5465-cb3f-44ae-a7bc-34ca246641a2 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Waiting for the task: (returnval){ [ 765.134207] env[63531]: value = "task-1117692" [ 765.134207] env[63531]: _type = "Task" [ 765.134207] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.142792] env[63531]: DEBUG oslo_vmware.api [None req-c52c5465-cb3f-44ae-a7bc-34ca246641a2 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': task-1117692, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.180014] env[63531]: DEBUG oslo_concurrency.lockutils [None req-fdd04ada-cc1b-4e03-9008-da80c3e4e411 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Acquiring lock "f39c8807-1e15-41cf-899b-a1fbe0695d58" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 765.180014] env[63531]: DEBUG oslo_concurrency.lockutils [None req-fdd04ada-cc1b-4e03-9008-da80c3e4e411 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Lock "f39c8807-1e15-41cf-899b-a1fbe0695d58" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 765.180014] env[63531]: DEBUG oslo_concurrency.lockutils [None req-fdd04ada-cc1b-4e03-9008-da80c3e4e411 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Acquiring lock "f39c8807-1e15-41cf-899b-a1fbe0695d58-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 765.180014] env[63531]: DEBUG oslo_concurrency.lockutils [None req-fdd04ada-cc1b-4e03-9008-da80c3e4e411 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Lock "f39c8807-1e15-41cf-899b-a1fbe0695d58-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 765.180014] env[63531]: DEBUG oslo_concurrency.lockutils [None req-fdd04ada-cc1b-4e03-9008-da80c3e4e411 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Lock "f39c8807-1e15-41cf-899b-a1fbe0695d58-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 765.182246] env[63531]: DEBUG nova.network.neutron [req-85da9566-24fa-4902-8ebb-de25e2e47a83 req-4e521678-f7cd-447d-bd32-97aa28a15146 service nova] [instance: dccb2335-4220-4570-861b-16ff2d8f5a85] Updated VIF entry in instance network info cache for port d69ccbf2-4502-4643-92b4-6fad825e76b4. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 765.182569] env[63531]: DEBUG nova.network.neutron [req-85da9566-24fa-4902-8ebb-de25e2e47a83 req-4e521678-f7cd-447d-bd32-97aa28a15146 service nova] [instance: dccb2335-4220-4570-861b-16ff2d8f5a85] Updating instance_info_cache with network_info: [{"id": "d69ccbf2-4502-4643-92b4-6fad825e76b4", "address": "fa:16:3e:43:06:ca", "network": {"id": "559568d9-859c-4dcf-9a3b-167f5dd3055c", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-251572816-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c4d75d29af5145bfb5140aabaeabf7e0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f880ac2e-d532-4f54-87bb-998a8d1bca78", "external-id": "nsx-vlan-transportzone-491", "segmentation_id": 491, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd69ccbf2-45", "ovs_interfaceid": "d69ccbf2-4502-4643-92b4-6fad825e76b4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 765.183836] env[63531]: INFO nova.compute.manager [None req-fdd04ada-cc1b-4e03-9008-da80c3e4e411 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: f39c8807-1e15-41cf-899b-a1fbe0695d58] Terminating instance [ 765.186044] env[63531]: DEBUG nova.compute.manager [None req-fdd04ada-cc1b-4e03-9008-da80c3e4e411 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: f39c8807-1e15-41cf-899b-a1fbe0695d58] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 765.186242] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-fdd04ada-cc1b-4e03-9008-da80c3e4e411 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: f39c8807-1e15-41cf-899b-a1fbe0695d58] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 765.187197] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89e9cc2f-01ba-48c2-942c-7a76b7db97ec {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.197483] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-fdd04ada-cc1b-4e03-9008-da80c3e4e411 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: f39c8807-1e15-41cf-899b-a1fbe0695d58] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 765.197682] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9eab5ba8-f12d-43d2-9e80-45157249a125 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.202219] env[63531]: DEBUG nova.scheduler.client.report [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Updated inventory for provider 9996a314-a62e-4282-8252-f387e7e3fabb with generation 71 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 765.202219] env[63531]: DEBUG nova.compute.provider_tree [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Updating resource provider 9996a314-a62e-4282-8252-f387e7e3fabb generation from 71 to 72 during operation: update_inventory {{(pid=63531) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 765.202219] env[63531]: DEBUG nova.compute.provider_tree [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Updating inventory in ProviderTree for provider 9996a314-a62e-4282-8252-f387e7e3fabb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 765.210361] env[63531]: DEBUG oslo_vmware.api [None req-fdd04ada-cc1b-4e03-9008-da80c3e4e411 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Waiting for the task: (returnval){ [ 765.210361] env[63531]: value = "task-1117693" [ 765.210361] env[63531]: _type = "Task" [ 765.210361] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.224046] env[63531]: DEBUG oslo_vmware.api [None req-fdd04ada-cc1b-4e03-9008-da80c3e4e411 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': task-1117693, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.257457] env[63531]: DEBUG oslo_vmware.api [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52d96786-a77c-c1ae-1129-5e0f6ff6c104, 'name': SearchDatastore_Task, 'duration_secs': 0.020063} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.258220] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 765.258485] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] 29dd6de3-2765-408c-acc0-da47e5e0a977/29dd6de3-2765-408c-acc0-da47e5e0a977.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 765.258752] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c6e1ace1-7045-4480-99ce-e40f87cc1c1b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.265518] env[63531]: DEBUG oslo_vmware.api [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 765.265518] env[63531]: value = "task-1117694" [ 765.265518] env[63531]: _type = "Task" [ 765.265518] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.274663] env[63531]: DEBUG oslo_vmware.api [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1117694, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.294833] env[63531]: DEBUG oslo_vmware.api [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Task: {'id': task-1117688, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.532343} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.295088] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] e1693594-d8bb-4002-b71c-7013efb81d16/e1693594-d8bb-4002-b71c-7013efb81d16.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 765.295305] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 765.295643] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b09cbc1a-e42a-42f7-b63d-e86f18bfdc5d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.304524] env[63531]: DEBUG oslo_vmware.api [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Waiting for the task: (returnval){ [ 765.304524] env[63531]: value = "task-1117695" [ 765.304524] env[63531]: _type = "Task" [ 765.304524] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.312756] env[63531]: DEBUG oslo_vmware.api [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Task: {'id': task-1117695, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.360059] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117691, 'name': CreateVM_Task, 'duration_secs': 0.406019} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.360059] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dccb2335-4220-4570-861b-16ff2d8f5a85] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 765.360787] env[63531]: DEBUG oslo_concurrency.lockutils [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 765.360954] env[63531]: DEBUG oslo_concurrency.lockutils [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.361293] env[63531]: DEBUG oslo_concurrency.lockutils [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 765.361580] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9fdb3db0-2bd7-49cb-a490-6dfc32e1ed53 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.366886] env[63531]: DEBUG oslo_vmware.api [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Waiting for the task: (returnval){ [ 765.366886] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52f58dcc-7172-e92a-3bd7-43a66351db32" [ 765.366886] env[63531]: _type = "Task" [ 765.366886] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.375250] env[63531]: DEBUG oslo_vmware.api [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52f58dcc-7172-e92a-3bd7-43a66351db32, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.643990] env[63531]: DEBUG oslo_vmware.api [None req-c52c5465-cb3f-44ae-a7bc-34ca246641a2 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': task-1117692, 'name': PowerOffVM_Task, 'duration_secs': 0.205196} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.644367] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-c52c5465-cb3f-44ae-a7bc-34ca246641a2 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: 645abb7a-cf73-4fb0-a9a0-49205f060d6f] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 765.644442] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c52c5465-cb3f-44ae-a7bc-34ca246641a2 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: 645abb7a-cf73-4fb0-a9a0-49205f060d6f] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 765.644734] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1eaa8df0-9a61-4659-a388-6cd8cb241e8b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.686389] env[63531]: DEBUG oslo_concurrency.lockutils [req-85da9566-24fa-4902-8ebb-de25e2e47a83 req-4e521678-f7cd-447d-bd32-97aa28a15146 service nova] Releasing lock "refresh_cache-dccb2335-4220-4570-861b-16ff2d8f5a85" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 765.710040] env[63531]: DEBUG oslo_concurrency.lockutils [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.686s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 765.710362] env[63531]: DEBUG nova.compute.manager [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 765.713721] env[63531]: DEBUG oslo_concurrency.lockutils [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 25.633s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 765.728317] env[63531]: DEBUG oslo_vmware.api [None req-fdd04ada-cc1b-4e03-9008-da80c3e4e411 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': task-1117693, 'name': PowerOffVM_Task, 'duration_secs': 0.19334} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.730786] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-fdd04ada-cc1b-4e03-9008-da80c3e4e411 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: f39c8807-1e15-41cf-899b-a1fbe0695d58] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 765.730786] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-fdd04ada-cc1b-4e03-9008-da80c3e4e411 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: f39c8807-1e15-41cf-899b-a1fbe0695d58] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 765.730786] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-771fed64-9a0d-48e6-b70f-e489e8092ae9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.743271] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c52c5465-cb3f-44ae-a7bc-34ca246641a2 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: 645abb7a-cf73-4fb0-a9a0-49205f060d6f] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 765.743436] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c52c5465-cb3f-44ae-a7bc-34ca246641a2 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: 645abb7a-cf73-4fb0-a9a0-49205f060d6f] Deleting contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 765.743621] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-c52c5465-cb3f-44ae-a7bc-34ca246641a2 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Deleting the datastore file [datastore2] 645abb7a-cf73-4fb0-a9a0-49205f060d6f {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 765.744181] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-61b44922-15a4-46d6-ad2a-c5b775d61e4e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.751437] env[63531]: DEBUG oslo_vmware.api [None req-c52c5465-cb3f-44ae-a7bc-34ca246641a2 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Waiting for the task: (returnval){ [ 765.751437] env[63531]: value = "task-1117698" [ 765.751437] env[63531]: _type = "Task" [ 765.751437] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.760924] env[63531]: DEBUG oslo_vmware.api [None req-c52c5465-cb3f-44ae-a7bc-34ca246641a2 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': task-1117698, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.775251] env[63531]: DEBUG oslo_vmware.api [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1117694, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.801284] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-fdd04ada-cc1b-4e03-9008-da80c3e4e411 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: f39c8807-1e15-41cf-899b-a1fbe0695d58] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 765.801531] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-fdd04ada-cc1b-4e03-9008-da80c3e4e411 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: f39c8807-1e15-41cf-899b-a1fbe0695d58] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 765.801757] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-fdd04ada-cc1b-4e03-9008-da80c3e4e411 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Deleting the datastore file [datastore1] f39c8807-1e15-41cf-899b-a1fbe0695d58 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 765.802066] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5c4447a7-c19c-452d-aec3-c72ed3192bd9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.814544] env[63531]: DEBUG oslo_vmware.api [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Task: {'id': task-1117695, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.091643} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.815810] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 765.816154] env[63531]: DEBUG oslo_vmware.api [None req-fdd04ada-cc1b-4e03-9008-da80c3e4e411 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Waiting for the task: (returnval){ [ 765.816154] env[63531]: value = "task-1117699" [ 765.816154] env[63531]: _type = "Task" [ 765.816154] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.816927] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f10f3ac9-1fc5-43f4-87e1-4576be29464d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.828310] env[63531]: DEBUG oslo_vmware.api [None req-fdd04ada-cc1b-4e03-9008-da80c3e4e411 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': task-1117699, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.851543] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Reconfiguring VM instance instance-0000002e to attach disk [datastore1] e1693594-d8bb-4002-b71c-7013efb81d16/e1693594-d8bb-4002-b71c-7013efb81d16.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 765.851879] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-68854a89-8718-4ccb-9351-c14d141fbe1b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.872124] env[63531]: DEBUG oslo_vmware.api [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Waiting for the task: (returnval){ [ 765.872124] env[63531]: value = "task-1117700" [ 765.872124] env[63531]: _type = "Task" [ 765.872124] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.878988] env[63531]: DEBUG oslo_vmware.api [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52f58dcc-7172-e92a-3bd7-43a66351db32, 'name': SearchDatastore_Task, 'duration_secs': 0.010372} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.879658] env[63531]: DEBUG oslo_concurrency.lockutils [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 765.879914] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] [instance: dccb2335-4220-4570-861b-16ff2d8f5a85] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 765.880177] env[63531]: DEBUG oslo_concurrency.lockutils [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 765.880335] env[63531]: DEBUG oslo_concurrency.lockutils [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.880515] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 765.880784] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f9da6418-50ba-42c0-8ef3-f29d667fb3f1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.885917] env[63531]: DEBUG oslo_vmware.api [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Task: {'id': task-1117700, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.893020] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 765.895889] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 765.895889] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ae319f6e-10ae-4dc9-acc0-9db9a491b087 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.899428] env[63531]: DEBUG oslo_vmware.api [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Waiting for the task: (returnval){ [ 765.899428] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]520eb17e-a3fb-81de-f561-187a4929e3c3" [ 765.899428] env[63531]: _type = "Task" [ 765.899428] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.907547] env[63531]: DEBUG oslo_vmware.api [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]520eb17e-a3fb-81de-f561-187a4929e3c3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.222440] env[63531]: DEBUG nova.compute.utils [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 766.226731] env[63531]: INFO nova.compute.claims [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 766.231019] env[63531]: DEBUG nova.compute.manager [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 766.231019] env[63531]: DEBUG nova.network.neutron [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 766.269425] env[63531]: DEBUG oslo_vmware.api [None req-c52c5465-cb3f-44ae-a7bc-34ca246641a2 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': task-1117698, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.153648} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.269425] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-c52c5465-cb3f-44ae-a7bc-34ca246641a2 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 766.269425] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c52c5465-cb3f-44ae-a7bc-34ca246641a2 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: 645abb7a-cf73-4fb0-a9a0-49205f060d6f] Deleted contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 766.269425] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c52c5465-cb3f-44ae-a7bc-34ca246641a2 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: 645abb7a-cf73-4fb0-a9a0-49205f060d6f] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 766.269425] env[63531]: INFO nova.compute.manager [None req-c52c5465-cb3f-44ae-a7bc-34ca246641a2 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: 645abb7a-cf73-4fb0-a9a0-49205f060d6f] Took 1.15 seconds to destroy the instance on the hypervisor. [ 766.269425] env[63531]: DEBUG oslo.service.loopingcall [None req-c52c5465-cb3f-44ae-a7bc-34ca246641a2 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 766.269425] env[63531]: DEBUG nova.compute.manager [-] [instance: 645abb7a-cf73-4fb0-a9a0-49205f060d6f] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 766.269425] env[63531]: DEBUG nova.network.neutron [-] [instance: 645abb7a-cf73-4fb0-a9a0-49205f060d6f] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 766.276198] env[63531]: DEBUG oslo_vmware.api [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1117694, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.326250] env[63531]: DEBUG nova.policy [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0a501b2c64c04dcda3596c8a1bd3fa0d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'abdae960d43b47bc95ab41f265d2c85b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 766.332781] env[63531]: DEBUG oslo_vmware.api [None req-fdd04ada-cc1b-4e03-9008-da80c3e4e411 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Task: {'id': task-1117699, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.149336} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.333046] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-fdd04ada-cc1b-4e03-9008-da80c3e4e411 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 766.333233] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-fdd04ada-cc1b-4e03-9008-da80c3e4e411 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: f39c8807-1e15-41cf-899b-a1fbe0695d58] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 766.333403] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-fdd04ada-cc1b-4e03-9008-da80c3e4e411 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: f39c8807-1e15-41cf-899b-a1fbe0695d58] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 766.333569] env[63531]: INFO nova.compute.manager [None req-fdd04ada-cc1b-4e03-9008-da80c3e4e411 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] [instance: f39c8807-1e15-41cf-899b-a1fbe0695d58] Took 1.15 seconds to destroy the instance on the hypervisor. [ 766.333800] env[63531]: DEBUG oslo.service.loopingcall [None req-fdd04ada-cc1b-4e03-9008-da80c3e4e411 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 766.335939] env[63531]: DEBUG nova.compute.manager [-] [instance: f39c8807-1e15-41cf-899b-a1fbe0695d58] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 766.335939] env[63531]: DEBUG nova.network.neutron [-] [instance: f39c8807-1e15-41cf-899b-a1fbe0695d58] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 766.382473] env[63531]: DEBUG oslo_vmware.api [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Task: {'id': task-1117700, 'name': ReconfigVM_Task, 'duration_secs': 0.30087} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.382739] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Reconfigured VM instance instance-0000002e to attach disk [datastore1] e1693594-d8bb-4002-b71c-7013efb81d16/e1693594-d8bb-4002-b71c-7013efb81d16.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 766.383412] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-38593358-d905-4597-b58e-21ff3040e793 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.391209] env[63531]: DEBUG oslo_vmware.api [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Waiting for the task: (returnval){ [ 766.391209] env[63531]: value = "task-1117701" [ 766.391209] env[63531]: _type = "Task" [ 766.391209] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.400877] env[63531]: DEBUG oslo_vmware.api [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Task: {'id': task-1117701, 'name': Rename_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.409265] env[63531]: DEBUG oslo_vmware.api [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]520eb17e-a3fb-81de-f561-187a4929e3c3, 'name': SearchDatastore_Task, 'duration_secs': 0.012491} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.410099] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-796f3ea7-f3aa-4c9c-944b-191ddbca8f59 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.415304] env[63531]: DEBUG oslo_vmware.api [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Waiting for the task: (returnval){ [ 766.415304] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5260b498-017e-4b6c-9bc0-c0c37ffee847" [ 766.415304] env[63531]: _type = "Task" [ 766.415304] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.422876] env[63531]: DEBUG oslo_vmware.api [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5260b498-017e-4b6c-9bc0-c0c37ffee847, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.732069] env[63531]: DEBUG nova.compute.manager [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 766.738124] env[63531]: INFO nova.compute.resource_tracker [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Updating resource usage from migration 97923ce2-3d93-4ee9-a54b-c15bc8e3c7e3 [ 766.775621] env[63531]: DEBUG nova.compute.manager [req-9b5dcaab-8356-4b9a-ae51-ae83209ce02f req-7e072856-28d1-41ad-8148-d97d4d80fdd4 service nova] [instance: 645abb7a-cf73-4fb0-a9a0-49205f060d6f] Received event network-vif-deleted-7d4594d0-2a5d-4a5c-9f4c-61bbe9731787 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 766.776182] env[63531]: INFO nova.compute.manager [req-9b5dcaab-8356-4b9a-ae51-ae83209ce02f req-7e072856-28d1-41ad-8148-d97d4d80fdd4 service nova] [instance: 645abb7a-cf73-4fb0-a9a0-49205f060d6f] Neutron deleted interface 7d4594d0-2a5d-4a5c-9f4c-61bbe9731787; detaching it from the instance and deleting it from the info cache [ 766.776416] env[63531]: DEBUG nova.network.neutron [req-9b5dcaab-8356-4b9a-ae51-ae83209ce02f req-7e072856-28d1-41ad-8148-d97d4d80fdd4 service nova] [instance: 645abb7a-cf73-4fb0-a9a0-49205f060d6f] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.789961] env[63531]: DEBUG oslo_vmware.api [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1117694, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.492932} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.790313] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] 29dd6de3-2765-408c-acc0-da47e5e0a977/29dd6de3-2765-408c-acc0-da47e5e0a977.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 766.791277] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 766.791277] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b21f49dc-838b-4754-8024-dc21e04c81c3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.800111] env[63531]: DEBUG oslo_vmware.api [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 766.800111] env[63531]: value = "task-1117702" [ 766.800111] env[63531]: _type = "Task" [ 766.800111] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.811285] env[63531]: DEBUG oslo_vmware.api [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1117702, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.901916] env[63531]: DEBUG oslo_vmware.api [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Task: {'id': task-1117701, 'name': Rename_Task, 'duration_secs': 0.234333} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.902319] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 766.902597] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-55d8ec73-47ac-48c2-a356-daff5f3120b3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.909373] env[63531]: DEBUG oslo_vmware.api [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Waiting for the task: (returnval){ [ 766.909373] env[63531]: value = "task-1117703" [ 766.909373] env[63531]: _type = "Task" [ 766.909373] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.922034] env[63531]: DEBUG oslo_vmware.api [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Task: {'id': task-1117703, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.929806] env[63531]: DEBUG oslo_vmware.api [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5260b498-017e-4b6c-9bc0-c0c37ffee847, 'name': SearchDatastore_Task, 'duration_secs': 0.023986} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.930073] env[63531]: DEBUG oslo_concurrency.lockutils [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 766.930337] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] dccb2335-4220-4570-861b-16ff2d8f5a85/dccb2335-4220-4570-861b-16ff2d8f5a85.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 766.930580] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3944738d-c7b1-46fb-9bf1-6d1a93cf754f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.935988] env[63531]: DEBUG oslo_vmware.api [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Waiting for the task: (returnval){ [ 766.935988] env[63531]: value = "task-1117704" [ 766.935988] env[63531]: _type = "Task" [ 766.935988] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.945801] env[63531]: DEBUG oslo_vmware.api [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Task: {'id': task-1117704, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.010195] env[63531]: DEBUG nova.network.neutron [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Successfully created port: 76060df9-1ff6-4c64-86ef-a8e0d4081cfd {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 767.144740] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e6acbc2-35bc-481b-9ac7-0000d5cbbd1f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.147782] env[63531]: DEBUG nova.network.neutron [-] [instance: 645abb7a-cf73-4fb0-a9a0-49205f060d6f] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.154860] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1723b0f7-0aee-4d8f-875a-f30033cb1bd3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.197196] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-006467c5-2e1b-4c71-adb6-58f4f6c73ed2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.208183] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea78bb9e-b59d-4437-975d-1c48a7eb9832 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.212860] env[63531]: DEBUG nova.network.neutron [-] [instance: f39c8807-1e15-41cf-899b-a1fbe0695d58] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.230344] env[63531]: DEBUG nova.compute.provider_tree [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 767.285274] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f4c88d4f-cda6-4bcf-a40f-40a4ac605a3a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.294519] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e54daa0-32f3-4708-9124-45f7706bfaca {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.318067] env[63531]: DEBUG oslo_vmware.api [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1117702, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068755} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.318351] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 767.319455] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65d31ff6-ac37-4ec5-a0b1-55db746f302a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.330766] env[63531]: DEBUG nova.compute.manager [req-9b5dcaab-8356-4b9a-ae51-ae83209ce02f req-7e072856-28d1-41ad-8148-d97d4d80fdd4 service nova] [instance: 645abb7a-cf73-4fb0-a9a0-49205f060d6f] Detach interface failed, port_id=7d4594d0-2a5d-4a5c-9f4c-61bbe9731787, reason: Instance 645abb7a-cf73-4fb0-a9a0-49205f060d6f could not be found. {{(pid=63531) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 767.356864] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Reconfiguring VM instance instance-0000002f to attach disk [datastore2] 29dd6de3-2765-408c-acc0-da47e5e0a977/29dd6de3-2765-408c-acc0-da47e5e0a977.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 767.357232] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b3ff2bdd-1bd0-440a-81f3-a302dc55c0c6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.378087] env[63531]: DEBUG oslo_vmware.api [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 767.378087] env[63531]: value = "task-1117705" [ 767.378087] env[63531]: _type = "Task" [ 767.378087] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.388723] env[63531]: DEBUG oslo_vmware.api [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1117705, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.421726] env[63531]: DEBUG oslo_vmware.api [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Task: {'id': task-1117703, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.448501] env[63531]: DEBUG oslo_vmware.api [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Task: {'id': task-1117704, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.506584} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.448501] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] dccb2335-4220-4570-861b-16ff2d8f5a85/dccb2335-4220-4570-861b-16ff2d8f5a85.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 767.448734] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] [instance: dccb2335-4220-4570-861b-16ff2d8f5a85] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 767.448949] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f4feef8d-535d-4636-9917-b3ce28a5a660 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.455878] env[63531]: DEBUG oslo_vmware.api [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Waiting for the task: (returnval){ [ 767.455878] env[63531]: value = "task-1117706" [ 767.455878] env[63531]: _type = "Task" [ 767.455878] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.465716] env[63531]: DEBUG oslo_vmware.api [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Task: {'id': task-1117706, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.651296] env[63531]: INFO nova.compute.manager [-] [instance: 645abb7a-cf73-4fb0-a9a0-49205f060d6f] Took 1.39 seconds to deallocate network for instance. [ 767.715501] env[63531]: INFO nova.compute.manager [-] [instance: f39c8807-1e15-41cf-899b-a1fbe0695d58] Took 1.38 seconds to deallocate network for instance. [ 767.729574] env[63531]: DEBUG nova.scheduler.client.report [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 767.746610] env[63531]: DEBUG nova.compute.manager [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 767.772879] env[63531]: DEBUG nova.virt.hardware [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 767.773053] env[63531]: DEBUG nova.virt.hardware [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 767.773292] env[63531]: DEBUG nova.virt.hardware [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 767.773489] env[63531]: DEBUG nova.virt.hardware [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 767.773637] env[63531]: DEBUG nova.virt.hardware [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 767.773782] env[63531]: DEBUG nova.virt.hardware [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 767.773987] env[63531]: DEBUG nova.virt.hardware [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 767.774164] env[63531]: DEBUG nova.virt.hardware [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 767.774329] env[63531]: DEBUG nova.virt.hardware [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 767.774564] env[63531]: DEBUG nova.virt.hardware [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 767.774657] env[63531]: DEBUG nova.virt.hardware [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 767.775507] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f78a91e-abc5-46bf-9c3c-d5f2f0a83f41 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.783478] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d021fa7-163b-4abd-b009-75651676a25d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.887944] env[63531]: DEBUG oslo_vmware.api [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1117705, 'name': ReconfigVM_Task, 'duration_secs': 0.33671} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.889105] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Reconfigured VM instance instance-0000002f to attach disk [datastore2] 29dd6de3-2765-408c-acc0-da47e5e0a977/29dd6de3-2765-408c-acc0-da47e5e0a977.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 767.889105] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-606318a2-a2fd-4a62-a7fd-a7b925de39f0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.895607] env[63531]: DEBUG oslo_vmware.api [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 767.895607] env[63531]: value = "task-1117707" [ 767.895607] env[63531]: _type = "Task" [ 767.895607] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.903649] env[63531]: DEBUG oslo_vmware.api [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1117707, 'name': Rename_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.920115] env[63531]: DEBUG oslo_vmware.api [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Task: {'id': task-1117703, 'name': PowerOnVM_Task, 'duration_secs': 0.564672} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.920449] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 767.920697] env[63531]: INFO nova.compute.manager [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Took 17.37 seconds to spawn the instance on the hypervisor. [ 767.920916] env[63531]: DEBUG nova.compute.manager [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 767.921714] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c41b7f5e-a9d8-4b44-93c8-b86d1f04ff34 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.965879] env[63531]: DEBUG oslo_vmware.api [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Task: {'id': task-1117706, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065867} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.966178] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] [instance: dccb2335-4220-4570-861b-16ff2d8f5a85] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 767.966936] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2eff9735-a553-4995-996a-89b777c66323 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.989792] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] [instance: dccb2335-4220-4570-861b-16ff2d8f5a85] Reconfiguring VM instance instance-00000030 to attach disk [datastore1] dccb2335-4220-4570-861b-16ff2d8f5a85/dccb2335-4220-4570-861b-16ff2d8f5a85.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 767.990333] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-89d2f8ff-2c4a-406d-be02-b7ace0d33312 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.009809] env[63531]: DEBUG oslo_vmware.api [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Waiting for the task: (returnval){ [ 768.009809] env[63531]: value = "task-1117708" [ 768.009809] env[63531]: _type = "Task" [ 768.009809] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.017235] env[63531]: DEBUG oslo_vmware.api [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Task: {'id': task-1117708, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.159089] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c52c5465-cb3f-44ae-a7bc-34ca246641a2 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 768.225934] env[63531]: DEBUG oslo_concurrency.lockutils [None req-fdd04ada-cc1b-4e03-9008-da80c3e4e411 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 768.233924] env[63531]: DEBUG oslo_concurrency.lockutils [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.520s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.234221] env[63531]: INFO nova.compute.manager [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Migrating [ 768.234449] env[63531]: DEBUG oslo_concurrency.lockutils [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 768.234621] env[63531]: DEBUG oslo_concurrency.lockutils [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Acquired lock "compute-rpcapi-router" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.236307] env[63531]: DEBUG oslo_concurrency.lockutils [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.329s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 768.238194] env[63531]: INFO nova.compute.claims [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 768.405402] env[63531]: DEBUG oslo_vmware.api [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1117707, 'name': Rename_Task, 'duration_secs': 0.325541} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.405726] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 768.405994] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-49db3125-906c-4dd6-93b0-fe5fb045f053 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.412114] env[63531]: DEBUG oslo_vmware.api [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 768.412114] env[63531]: value = "task-1117709" [ 768.412114] env[63531]: _type = "Task" [ 768.412114] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.419578] env[63531]: DEBUG oslo_vmware.api [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1117709, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.439566] env[63531]: INFO nova.compute.manager [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Took 41.07 seconds to build instance. [ 768.519741] env[63531]: DEBUG oslo_vmware.api [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Task: {'id': task-1117708, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.737298] env[63531]: INFO nova.compute.rpcapi [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Automatically selected compute RPC version 6.3 from minimum service version 67 [ 768.738023] env[63531]: DEBUG oslo_concurrency.lockutils [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Releasing lock "compute-rpcapi-router" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 768.922223] env[63531]: DEBUG oslo_vmware.api [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1117709, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.941846] env[63531]: DEBUG nova.compute.manager [req-ad2a5da1-a8b9-4d42-9c8a-3e8147958127 req-e71ed52a-372b-4ca7-bcd7-9af241aef871 service nova] [instance: f39c8807-1e15-41cf-899b-a1fbe0695d58] Received event network-vif-deleted-61792095-541d-4d6e-b33c-c35d459fe790 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 768.942116] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a7b3f15f-3821-48d9-86aa-933796782959 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Lock "e1693594-d8bb-4002-b71c-7013efb81d16" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 139.621s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.021839] env[63531]: DEBUG oslo_vmware.api [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Task: {'id': task-1117708, 'name': ReconfigVM_Task, 'duration_secs': 0.845399} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.021839] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] [instance: dccb2335-4220-4570-861b-16ff2d8f5a85] Reconfigured VM instance instance-00000030 to attach disk [datastore1] dccb2335-4220-4570-861b-16ff2d8f5a85/dccb2335-4220-4570-861b-16ff2d8f5a85.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 769.022522] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4421fafc-03a7-4fad-bcd0-341679d26e13 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.028752] env[63531]: DEBUG oslo_vmware.api [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Waiting for the task: (returnval){ [ 769.028752] env[63531]: value = "task-1117710" [ 769.028752] env[63531]: _type = "Task" [ 769.028752] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.036826] env[63531]: DEBUG oslo_vmware.api [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Task: {'id': task-1117710, 'name': Rename_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.120906] env[63531]: DEBUG nova.network.neutron [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Successfully updated port: 76060df9-1ff6-4c64-86ef-a8e0d4081cfd {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 769.142234] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-659ac8bf-7fcf-4e55-b556-d76ffef83969 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.154989] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b144774-2c6a-476e-852f-7577ec3ca372 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.160693] env[63531]: DEBUG nova.compute.manager [req-2cc20592-97ec-410d-a51c-5844bb9b88db req-6dbe7f23-81b2-407c-bdd0-e26c8a3ac456 service nova] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Received event network-vif-plugged-76060df9-1ff6-4c64-86ef-a8e0d4081cfd {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 769.160693] env[63531]: DEBUG oslo_concurrency.lockutils [req-2cc20592-97ec-410d-a51c-5844bb9b88db req-6dbe7f23-81b2-407c-bdd0-e26c8a3ac456 service nova] Acquiring lock "8dec0cac-defc-42f2-ab56-6b3ae60ad858-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 769.160693] env[63531]: DEBUG oslo_concurrency.lockutils [req-2cc20592-97ec-410d-a51c-5844bb9b88db req-6dbe7f23-81b2-407c-bdd0-e26c8a3ac456 service nova] Lock "8dec0cac-defc-42f2-ab56-6b3ae60ad858-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 769.160693] env[63531]: DEBUG oslo_concurrency.lockutils [req-2cc20592-97ec-410d-a51c-5844bb9b88db req-6dbe7f23-81b2-407c-bdd0-e26c8a3ac456 service nova] Lock "8dec0cac-defc-42f2-ab56-6b3ae60ad858-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.160693] env[63531]: DEBUG nova.compute.manager [req-2cc20592-97ec-410d-a51c-5844bb9b88db req-6dbe7f23-81b2-407c-bdd0-e26c8a3ac456 service nova] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] No waiting events found dispatching network-vif-plugged-76060df9-1ff6-4c64-86ef-a8e0d4081cfd {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 769.160693] env[63531]: WARNING nova.compute.manager [req-2cc20592-97ec-410d-a51c-5844bb9b88db req-6dbe7f23-81b2-407c-bdd0-e26c8a3ac456 service nova] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Received unexpected event network-vif-plugged-76060df9-1ff6-4c64-86ef-a8e0d4081cfd for instance with vm_state building and task_state spawning. [ 769.188344] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cee0d8ac-f674-44f0-934c-d014e64ec622 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.195859] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41952d5d-e2ea-4612-b2c3-4d285b77875a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.209581] env[63531]: DEBUG nova.compute.provider_tree [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Updating inventory in ProviderTree for provider 9996a314-a62e-4282-8252-f387e7e3fabb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 769.272684] env[63531]: DEBUG oslo_concurrency.lockutils [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Acquiring lock "refresh_cache-5e9042b3-4dc2-4fa3-a664-c4b49a22e400" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 769.272684] env[63531]: DEBUG oslo_concurrency.lockutils [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Acquired lock "refresh_cache-5e9042b3-4dc2-4fa3-a664-c4b49a22e400" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.272684] env[63531]: DEBUG nova.network.neutron [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 769.425071] env[63531]: DEBUG oslo_vmware.api [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1117709, 'name': PowerOnVM_Task, 'duration_secs': 0.518713} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.425320] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 769.425518] env[63531]: INFO nova.compute.manager [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Took 10.03 seconds to spawn the instance on the hypervisor. [ 769.425760] env[63531]: DEBUG nova.compute.manager [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 769.426530] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44047e02-0b21-4892-b545-f1bc7765ad18 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.444784] env[63531]: DEBUG nova.compute.manager [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 769.538149] env[63531]: DEBUG oslo_vmware.api [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Task: {'id': task-1117710, 'name': Rename_Task, 'duration_secs': 0.217742} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.538433] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] [instance: dccb2335-4220-4570-861b-16ff2d8f5a85] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 769.539026] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-83e0c945-21cc-4525-9fe7-9de751aec24d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.544361] env[63531]: DEBUG oslo_vmware.api [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Waiting for the task: (returnval){ [ 769.544361] env[63531]: value = "task-1117711" [ 769.544361] env[63531]: _type = "Task" [ 769.544361] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.551967] env[63531]: DEBUG oslo_vmware.api [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Task: {'id': task-1117711, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.624570] env[63531]: DEBUG oslo_concurrency.lockutils [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquiring lock "refresh_cache-8dec0cac-defc-42f2-ab56-6b3ae60ad858" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 769.624570] env[63531]: DEBUG oslo_concurrency.lockutils [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquired lock "refresh_cache-8dec0cac-defc-42f2-ab56-6b3ae60ad858" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.624570] env[63531]: DEBUG nova.network.neutron [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 769.743581] env[63531]: DEBUG nova.scheduler.client.report [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Updated inventory for provider 9996a314-a62e-4282-8252-f387e7e3fabb with generation 72 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 769.743848] env[63531]: DEBUG nova.compute.provider_tree [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Updating resource provider 9996a314-a62e-4282-8252-f387e7e3fabb generation from 72 to 73 during operation: update_inventory {{(pid=63531) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 769.744042] env[63531]: DEBUG nova.compute.provider_tree [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Updating inventory in ProviderTree for provider 9996a314-a62e-4282-8252-f387e7e3fabb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 769.887558] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7385c1eb-e474-478e-a87d-01f2ddd50615 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Acquiring lock "e1693594-d8bb-4002-b71c-7013efb81d16" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 769.887948] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7385c1eb-e474-478e-a87d-01f2ddd50615 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Lock "e1693594-d8bb-4002-b71c-7013efb81d16" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 769.888219] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7385c1eb-e474-478e-a87d-01f2ddd50615 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Acquiring lock "e1693594-d8bb-4002-b71c-7013efb81d16-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 769.888346] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7385c1eb-e474-478e-a87d-01f2ddd50615 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Lock "e1693594-d8bb-4002-b71c-7013efb81d16-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 769.888529] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7385c1eb-e474-478e-a87d-01f2ddd50615 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Lock "e1693594-d8bb-4002-b71c-7013efb81d16-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.893413] env[63531]: INFO nova.compute.manager [None req-7385c1eb-e474-478e-a87d-01f2ddd50615 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Terminating instance [ 769.895350] env[63531]: DEBUG nova.compute.manager [None req-7385c1eb-e474-478e-a87d-01f2ddd50615 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 769.895601] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-7385c1eb-e474-478e-a87d-01f2ddd50615 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 769.896466] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b0eda14-e2d7-4892-ab48-346aab9416ae {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.904297] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-7385c1eb-e474-478e-a87d-01f2ddd50615 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 769.904500] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-78f10d99-0ea3-4677-aa6a-d28c6ed58a51 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.911298] env[63531]: DEBUG oslo_vmware.api [None req-7385c1eb-e474-478e-a87d-01f2ddd50615 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Waiting for the task: (returnval){ [ 769.911298] env[63531]: value = "task-1117712" [ 769.911298] env[63531]: _type = "Task" [ 769.911298] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.918616] env[63531]: DEBUG oslo_vmware.api [None req-7385c1eb-e474-478e-a87d-01f2ddd50615 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Task: {'id': task-1117712, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.946193] env[63531]: INFO nova.compute.manager [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Took 36.23 seconds to build instance. [ 769.976396] env[63531]: DEBUG oslo_concurrency.lockutils [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 770.054653] env[63531]: DEBUG oslo_vmware.api [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Task: {'id': task-1117711, 'name': PowerOnVM_Task} progress is 90%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.139245] env[63531]: DEBUG nova.network.neutron [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Updating instance_info_cache with network_info: [{"id": "f178c862-e753-40d4-b7cd-3eabe7eda39e", "address": "fa:16:3e:a6:d3:d2", "network": {"id": "5439232a-ec6d-4184-a9f1-f7be5c1526bc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.35", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "54fbef8a6ee0488f820706d2036e9c99", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f267bcdd-0daa-4337-9709-5fc060c267d8", "external-id": "nsx-vlan-transportzone-308", "segmentation_id": 308, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf178c862-e7", "ovs_interfaceid": "f178c862-e753-40d4-b7cd-3eabe7eda39e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.170921] env[63531]: DEBUG nova.network.neutron [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 770.248696] env[63531]: DEBUG oslo_concurrency.lockutils [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.012s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 770.249312] env[63531]: DEBUG nova.compute.manager [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 770.252158] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.198s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 770.253524] env[63531]: INFO nova.compute.claims [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 770.319743] env[63531]: DEBUG nova.network.neutron [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Updating instance_info_cache with network_info: [{"id": "76060df9-1ff6-4c64-86ef-a8e0d4081cfd", "address": "fa:16:3e:52:f5:96", "network": {"id": "eac8786d-bc67-4916-8483-b53d9a25efb8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1604989690-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "abdae960d43b47bc95ab41f265d2c85b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db068f71-08cc-42d4-8ab6-17134c1585e5", "external-id": "nsx-vlan-transportzone-721", "segmentation_id": 721, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap76060df9-1f", "ovs_interfaceid": "76060df9-1ff6-4c64-86ef-a8e0d4081cfd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.422091] env[63531]: DEBUG oslo_vmware.api [None req-7385c1eb-e474-478e-a87d-01f2ddd50615 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Task: {'id': task-1117712, 'name': PowerOffVM_Task, 'duration_secs': 0.179387} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.422342] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-7385c1eb-e474-478e-a87d-01f2ddd50615 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 770.422621] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-7385c1eb-e474-478e-a87d-01f2ddd50615 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 770.422743] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-806a1539-8dd2-4363-b1b2-2fedb0f15afb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.448866] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e85016ba-5bc0-4835-9df3-01cd2be0fb8e tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "29dd6de3-2765-408c-acc0-da47e5e0a977" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 133.319s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 770.554693] env[63531]: DEBUG oslo_vmware.api [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Task: {'id': task-1117711, 'name': PowerOnVM_Task, 'duration_secs': 0.584141} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.554988] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] [instance: dccb2335-4220-4570-861b-16ff2d8f5a85] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 770.555323] env[63531]: INFO nova.compute.manager [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] [instance: dccb2335-4220-4570-861b-16ff2d8f5a85] Took 8.57 seconds to spawn the instance on the hypervisor. [ 770.555511] env[63531]: DEBUG nova.compute.manager [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] [instance: dccb2335-4220-4570-861b-16ff2d8f5a85] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 770.556321] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e53b439-fba0-4f9e-a9e0-9f2dbdd944e3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.561558] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-7385c1eb-e474-478e-a87d-01f2ddd50615 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 770.561755] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-7385c1eb-e474-478e-a87d-01f2ddd50615 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 770.561933] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-7385c1eb-e474-478e-a87d-01f2ddd50615 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Deleting the datastore file [datastore1] e1693594-d8bb-4002-b71c-7013efb81d16 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 770.562199] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ce5e94d5-6b66-4499-a600-e8781f3666fe {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.573617] env[63531]: DEBUG oslo_vmware.api [None req-7385c1eb-e474-478e-a87d-01f2ddd50615 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Waiting for the task: (returnval){ [ 770.573617] env[63531]: value = "task-1117714" [ 770.573617] env[63531]: _type = "Task" [ 770.573617] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.581669] env[63531]: DEBUG oslo_vmware.api [None req-7385c1eb-e474-478e-a87d-01f2ddd50615 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Task: {'id': task-1117714, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.643165] env[63531]: DEBUG oslo_concurrency.lockutils [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Releasing lock "refresh_cache-5e9042b3-4dc2-4fa3-a664-c4b49a22e400" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 770.764292] env[63531]: DEBUG nova.compute.utils [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 770.764292] env[63531]: DEBUG nova.compute.manager [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 770.764292] env[63531]: DEBUG nova.network.neutron [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 770.824753] env[63531]: DEBUG nova.policy [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd0112e02798343e98ad8a723f458cade', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '401b45e0117342589abdfd44bfcc1a47', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 770.827385] env[63531]: DEBUG oslo_concurrency.lockutils [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Releasing lock "refresh_cache-8dec0cac-defc-42f2-ab56-6b3ae60ad858" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 770.827713] env[63531]: DEBUG nova.compute.manager [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Instance network_info: |[{"id": "76060df9-1ff6-4c64-86ef-a8e0d4081cfd", "address": "fa:16:3e:52:f5:96", "network": {"id": "eac8786d-bc67-4916-8483-b53d9a25efb8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1604989690-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "abdae960d43b47bc95ab41f265d2c85b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db068f71-08cc-42d4-8ab6-17134c1585e5", "external-id": "nsx-vlan-transportzone-721", "segmentation_id": 721, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap76060df9-1f", "ovs_interfaceid": "76060df9-1ff6-4c64-86ef-a8e0d4081cfd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 770.828241] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:52:f5:96', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'db068f71-08cc-42d4-8ab6-17134c1585e5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '76060df9-1ff6-4c64-86ef-a8e0d4081cfd', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 770.837321] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Creating folder: Project (abdae960d43b47bc95ab41f265d2c85b). Parent ref: group-v244585. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 770.837966] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1cca482b-dfc9-4f59-a297-5ce9f2e9cc86 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.848037] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Created folder: Project (abdae960d43b47bc95ab41f265d2c85b) in parent group-v244585. [ 770.848161] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Creating folder: Instances. Parent ref: group-v244649. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 770.848395] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2832ae8b-b18a-40e9-99f9-7ce534d13d62 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.856962] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Created folder: Instances in parent group-v244649. [ 770.857228] env[63531]: DEBUG oslo.service.loopingcall [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 770.857397] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 770.860020] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c037c811-6bd8-4414-867c-965927c40ede {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.877370] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 770.877370] env[63531]: value = "task-1117717" [ 770.877370] env[63531]: _type = "Task" [ 770.877370] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.883442] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117717, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.952273] env[63531]: DEBUG nova.compute.manager [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 771.083902] env[63531]: INFO nova.compute.manager [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] [instance: dccb2335-4220-4570-861b-16ff2d8f5a85] Took 35.83 seconds to build instance. [ 771.091124] env[63531]: DEBUG oslo_vmware.api [None req-7385c1eb-e474-478e-a87d-01f2ddd50615 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Task: {'id': task-1117714, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.165581} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.091124] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-7385c1eb-e474-478e-a87d-01f2ddd50615 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 771.091124] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-7385c1eb-e474-478e-a87d-01f2ddd50615 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 771.091124] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-7385c1eb-e474-478e-a87d-01f2ddd50615 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 771.091124] env[63531]: INFO nova.compute.manager [None req-7385c1eb-e474-478e-a87d-01f2ddd50615 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Took 1.20 seconds to destroy the instance on the hypervisor. [ 771.091124] env[63531]: DEBUG oslo.service.loopingcall [None req-7385c1eb-e474-478e-a87d-01f2ddd50615 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 771.091124] env[63531]: DEBUG nova.compute.manager [-] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 771.091484] env[63531]: DEBUG nova.network.neutron [-] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 771.179775] env[63531]: DEBUG nova.network.neutron [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Successfully created port: f02115dd-c7b4-46f6-8057-20f0e4985b2e {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 771.272021] env[63531]: DEBUG nova.compute.manager [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 771.303015] env[63531]: DEBUG nova.compute.manager [req-8e6c56ef-13cd-4f8b-9ccd-559ea27728f9 req-748f5ab8-a2ea-451a-8af1-056b77c0e318 service nova] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Received event network-changed-76060df9-1ff6-4c64-86ef-a8e0d4081cfd {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 771.303208] env[63531]: DEBUG nova.compute.manager [req-8e6c56ef-13cd-4f8b-9ccd-559ea27728f9 req-748f5ab8-a2ea-451a-8af1-056b77c0e318 service nova] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Refreshing instance network info cache due to event network-changed-76060df9-1ff6-4c64-86ef-a8e0d4081cfd. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 771.303436] env[63531]: DEBUG oslo_concurrency.lockutils [req-8e6c56ef-13cd-4f8b-9ccd-559ea27728f9 req-748f5ab8-a2ea-451a-8af1-056b77c0e318 service nova] Acquiring lock "refresh_cache-8dec0cac-defc-42f2-ab56-6b3ae60ad858" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 771.303574] env[63531]: DEBUG oslo_concurrency.lockutils [req-8e6c56ef-13cd-4f8b-9ccd-559ea27728f9 req-748f5ab8-a2ea-451a-8af1-056b77c0e318 service nova] Acquired lock "refresh_cache-8dec0cac-defc-42f2-ab56-6b3ae60ad858" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 771.303736] env[63531]: DEBUG nova.network.neutron [req-8e6c56ef-13cd-4f8b-9ccd-559ea27728f9 req-748f5ab8-a2ea-451a-8af1-056b77c0e318 service nova] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Refreshing network info cache for port 76060df9-1ff6-4c64-86ef-a8e0d4081cfd {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 771.395719] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117717, 'name': CreateVM_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.476739] env[63531]: DEBUG oslo_concurrency.lockutils [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 771.586979] env[63531]: DEBUG oslo_concurrency.lockutils [None req-315ffb71-b305-414f-9df0-a3a7382f6243 tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Lock "dccb2335-4220-4570-861b-16ff2d8f5a85" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 123.081s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 772.395415] env[63531]: DEBUG nova.compute.manager [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 531e1852-0c67-4b4d-a0ca-749ac438e79a] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 772.406503] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46621620-a452-41c6-a3fb-20d822ca57f7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.414359] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117717, 'name': CreateVM_Task, 'duration_secs': 0.540268} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.427506] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 772.428469] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Updating instance '5e9042b3-4dc2-4fa3-a664-c4b49a22e400' progress to 0 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 772.435338] env[63531]: DEBUG oslo_concurrency.lockutils [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 772.435522] env[63531]: DEBUG oslo_concurrency.lockutils [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 772.436283] env[63531]: DEBUG oslo_concurrency.lockutils [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 772.439096] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c7aa0774-8fe4-403d-a4a7-2f52858f6b67 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.444901] env[63531]: DEBUG oslo_vmware.api [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 772.444901] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5261646e-4d1f-a1c6-697f-aecdfac77c52" [ 772.444901] env[63531]: _type = "Task" [ 772.444901] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.453613] env[63531]: DEBUG oslo_vmware.api [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5261646e-4d1f-a1c6-697f-aecdfac77c52, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.522430] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a07bdf49-ca27-4c5f-a64a-9e45866095aa {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.530157] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2f2f66a-9929-48fb-b543-62f01a84d5d7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.564253] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0d8c552-d8e6-4444-ad58-98e3b056aa89 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.571613] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efa7ba14-15ab-42bf-b928-096fb45e82a8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.587238] env[63531]: DEBUG nova.compute.provider_tree [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 772.732909] env[63531]: DEBUG nova.network.neutron [req-8e6c56ef-13cd-4f8b-9ccd-559ea27728f9 req-748f5ab8-a2ea-451a-8af1-056b77c0e318 service nova] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Updated VIF entry in instance network info cache for port 76060df9-1ff6-4c64-86ef-a8e0d4081cfd. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 772.732909] env[63531]: DEBUG nova.network.neutron [req-8e6c56ef-13cd-4f8b-9ccd-559ea27728f9 req-748f5ab8-a2ea-451a-8af1-056b77c0e318 service nova] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Updating instance_info_cache with network_info: [{"id": "76060df9-1ff6-4c64-86ef-a8e0d4081cfd", "address": "fa:16:3e:52:f5:96", "network": {"id": "eac8786d-bc67-4916-8483-b53d9a25efb8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1604989690-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "abdae960d43b47bc95ab41f265d2c85b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db068f71-08cc-42d4-8ab6-17134c1585e5", "external-id": "nsx-vlan-transportzone-721", "segmentation_id": 721, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap76060df9-1f", "ovs_interfaceid": "76060df9-1ff6-4c64-86ef-a8e0d4081cfd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 772.901527] env[63531]: DEBUG nova.compute.manager [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 772.939021] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 772.941909] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 772.942897] env[63531]: DEBUG oslo_concurrency.lockutils [None req-91e9db6d-db1f-484a-9006-e9fab2e8db3e tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Acquiring lock "dccb2335-4220-4570-861b-16ff2d8f5a85" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 772.943284] env[63531]: DEBUG oslo_concurrency.lockutils [None req-91e9db6d-db1f-484a-9006-e9fab2e8db3e tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Lock "dccb2335-4220-4570-861b-16ff2d8f5a85" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 772.944191] env[63531]: DEBUG oslo_concurrency.lockutils [None req-91e9db6d-db1f-484a-9006-e9fab2e8db3e tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Acquiring lock "dccb2335-4220-4570-861b-16ff2d8f5a85-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 772.944563] env[63531]: DEBUG oslo_concurrency.lockutils [None req-91e9db6d-db1f-484a-9006-e9fab2e8db3e tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Lock "dccb2335-4220-4570-861b-16ff2d8f5a85-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 772.944936] env[63531]: DEBUG oslo_concurrency.lockutils [None req-91e9db6d-db1f-484a-9006-e9fab2e8db3e tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Lock "dccb2335-4220-4570-861b-16ff2d8f5a85-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 772.946695] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e80a703f-5033-4742-b038-46a9ca433689 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.950491] env[63531]: DEBUG nova.virt.hardware [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 772.950874] env[63531]: DEBUG nova.virt.hardware [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 772.952023] env[63531]: DEBUG nova.virt.hardware [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 772.952023] env[63531]: DEBUG nova.virt.hardware [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 772.952023] env[63531]: DEBUG nova.virt.hardware [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 772.952023] env[63531]: DEBUG nova.virt.hardware [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 772.952023] env[63531]: DEBUG nova.virt.hardware [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 772.952023] env[63531]: DEBUG nova.virt.hardware [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 772.952645] env[63531]: DEBUG nova.virt.hardware [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 772.952936] env[63531]: DEBUG nova.virt.hardware [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 772.953258] env[63531]: DEBUG nova.virt.hardware [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 772.953836] env[63531]: INFO nova.compute.manager [None req-91e9db6d-db1f-484a-9006-e9fab2e8db3e tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] [instance: dccb2335-4220-4570-861b-16ff2d8f5a85] Terminating instance [ 772.957161] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-830dd802-501c-44f5-a3fb-83a4643db3df {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.960059] env[63531]: DEBUG nova.compute.manager [None req-91e9db6d-db1f-484a-9006-e9fab2e8db3e tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] [instance: dccb2335-4220-4570-861b-16ff2d8f5a85] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 772.960372] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-91e9db6d-db1f-484a-9006-e9fab2e8db3e tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] [instance: dccb2335-4220-4570-861b-16ff2d8f5a85] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 772.964603] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d6848f3-33f4-4e57-8c4a-438efbde43b5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.970183] env[63531]: DEBUG oslo_vmware.api [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Waiting for the task: (returnval){ [ 772.970183] env[63531]: value = "task-1117718" [ 772.970183] env[63531]: _type = "Task" [ 772.970183] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.983951] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4001c9b4-97fa-412e-a655-568354a5cfab {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.987959] env[63531]: DEBUG oslo_vmware.api [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5261646e-4d1f-a1c6-697f-aecdfac77c52, 'name': SearchDatastore_Task, 'duration_secs': 0.009446} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.989053] env[63531]: DEBUG oslo_concurrency.lockutils [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 772.989348] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 772.989607] env[63531]: DEBUG oslo_concurrency.lockutils [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 772.989758] env[63531]: DEBUG oslo_concurrency.lockutils [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 772.989964] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 772.993721] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4fdcf7da-b3c6-432c-b9e4-2bb6a8942b0c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.995801] env[63531]: DEBUG oslo_vmware.api [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Task: {'id': task-1117718, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.005642] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-91e9db6d-db1f-484a-9006-e9fab2e8db3e tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] [instance: dccb2335-4220-4570-861b-16ff2d8f5a85] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 773.006397] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a67f0a4a-de35-44c6-9f83-721f8d39d50e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.012475] env[63531]: DEBUG oslo_vmware.api [None req-91e9db6d-db1f-484a-9006-e9fab2e8db3e tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Waiting for the task: (returnval){ [ 773.012475] env[63531]: value = "task-1117719" [ 773.012475] env[63531]: _type = "Task" [ 773.012475] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.013544] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 773.013719] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 773.017034] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-139c1b8b-91b8-4a10-afb2-0adc796f221b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.022476] env[63531]: DEBUG oslo_vmware.api [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 773.022476] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]527bf48f-7636-87ee-13f5-add172db02c4" [ 773.022476] env[63531]: _type = "Task" [ 773.022476] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.025484] env[63531]: DEBUG oslo_vmware.api [None req-91e9db6d-db1f-484a-9006-e9fab2e8db3e tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Task: {'id': task-1117719, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.032959] env[63531]: DEBUG oslo_vmware.api [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]527bf48f-7636-87ee-13f5-add172db02c4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.091157] env[63531]: DEBUG nova.scheduler.client.report [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 773.136602] env[63531]: DEBUG nova.compute.manager [req-88dfc127-d6af-4b1c-a66a-406cf7e8016f req-a890cc5b-7671-4b0f-aeb6-40788afd3388 service nova] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Received event network-vif-plugged-f02115dd-c7b4-46f6-8057-20f0e4985b2e {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 773.136797] env[63531]: DEBUG oslo_concurrency.lockutils [req-88dfc127-d6af-4b1c-a66a-406cf7e8016f req-a890cc5b-7671-4b0f-aeb6-40788afd3388 service nova] Acquiring lock "2998fb3d-6677-4fb3-952a-c74d8c6b7b9f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 773.137129] env[63531]: DEBUG oslo_concurrency.lockutils [req-88dfc127-d6af-4b1c-a66a-406cf7e8016f req-a890cc5b-7671-4b0f-aeb6-40788afd3388 service nova] Lock "2998fb3d-6677-4fb3-952a-c74d8c6b7b9f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 773.137308] env[63531]: DEBUG oslo_concurrency.lockutils [req-88dfc127-d6af-4b1c-a66a-406cf7e8016f req-a890cc5b-7671-4b0f-aeb6-40788afd3388 service nova] Lock "2998fb3d-6677-4fb3-952a-c74d8c6b7b9f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 773.137473] env[63531]: DEBUG nova.compute.manager [req-88dfc127-d6af-4b1c-a66a-406cf7e8016f req-a890cc5b-7671-4b0f-aeb6-40788afd3388 service nova] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] No waiting events found dispatching network-vif-plugged-f02115dd-c7b4-46f6-8057-20f0e4985b2e {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 773.137635] env[63531]: WARNING nova.compute.manager [req-88dfc127-d6af-4b1c-a66a-406cf7e8016f req-a890cc5b-7671-4b0f-aeb6-40788afd3388 service nova] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Received unexpected event network-vif-plugged-f02115dd-c7b4-46f6-8057-20f0e4985b2e for instance with vm_state building and task_state spawning. [ 773.234693] env[63531]: DEBUG oslo_concurrency.lockutils [req-8e6c56ef-13cd-4f8b-9ccd-559ea27728f9 req-748f5ab8-a2ea-451a-8af1-056b77c0e318 service nova] Releasing lock "refresh_cache-8dec0cac-defc-42f2-ab56-6b3ae60ad858" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 773.236044] env[63531]: DEBUG nova.compute.manager [req-8e6c56ef-13cd-4f8b-9ccd-559ea27728f9 req-748f5ab8-a2ea-451a-8af1-056b77c0e318 service nova] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Received event network-changed-262abfb7-7bcf-40e6-9ce3-b502de45b3b4 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 773.236044] env[63531]: DEBUG nova.compute.manager [req-8e6c56ef-13cd-4f8b-9ccd-559ea27728f9 req-748f5ab8-a2ea-451a-8af1-056b77c0e318 service nova] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Refreshing instance network info cache due to event network-changed-262abfb7-7bcf-40e6-9ce3-b502de45b3b4. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 773.236044] env[63531]: DEBUG oslo_concurrency.lockutils [req-8e6c56ef-13cd-4f8b-9ccd-559ea27728f9 req-748f5ab8-a2ea-451a-8af1-056b77c0e318 service nova] Acquiring lock "refresh_cache-29dd6de3-2765-408c-acc0-da47e5e0a977" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 773.236044] env[63531]: DEBUG oslo_concurrency.lockutils [req-8e6c56ef-13cd-4f8b-9ccd-559ea27728f9 req-748f5ab8-a2ea-451a-8af1-056b77c0e318 service nova] Acquired lock "refresh_cache-29dd6de3-2765-408c-acc0-da47e5e0a977" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.236044] env[63531]: DEBUG nova.network.neutron [req-8e6c56ef-13cd-4f8b-9ccd-559ea27728f9 req-748f5ab8-a2ea-451a-8af1-056b77c0e318 service nova] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Refreshing network info cache for port 262abfb7-7bcf-40e6-9ce3-b502de45b3b4 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 773.327089] env[63531]: DEBUG nova.compute.manager [req-5db97274-8a7f-486e-a4b2-fa7562bd7370 req-43e20959-febe-45c0-8cc4-653716906ff2 service nova] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Received event network-vif-deleted-4da738f4-b2e9-40c4-9a02-43b6ed658cfa {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 773.327976] env[63531]: INFO nova.compute.manager [req-5db97274-8a7f-486e-a4b2-fa7562bd7370 req-43e20959-febe-45c0-8cc4-653716906ff2 service nova] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Neutron deleted interface 4da738f4-b2e9-40c4-9a02-43b6ed658cfa; detaching it from the instance and deleting it from the info cache [ 773.327976] env[63531]: DEBUG nova.network.neutron [req-5db97274-8a7f-486e-a4b2-fa7562bd7370 req-43e20959-febe-45c0-8cc4-653716906ff2 service nova] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Updating instance_info_cache with network_info: [{"id": "4c96ebbd-98fa-4d12-a388-b9d270b7aaae", "address": "fa:16:3e:97:f3:0b", "network": {"id": "1c410a92-1d7e-4b46-b7e0-7bcfd5dd7874", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1094381684", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.120", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bebde0e281d485ebdd70a1a3b613632", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d689fd7-f53e-4fd3-80d9-8d6b8fb7a164", "external-id": "nsx-vlan-transportzone-972", "segmentation_id": 972, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c96ebbd-98", "ovs_interfaceid": "4c96ebbd-98fa-4d12-a388-b9d270b7aaae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "5765492b-2d11-438b-b9a4-7514ce2ac73e", "address": "fa:16:3e:f7:f1:ba", "network": {"id": "485000a3-2bad-4c7d-b389-d4200550cc26", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-446289262", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.66", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "8bebde0e281d485ebdd70a1a3b613632", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5765492b-2d", "ovs_interfaceid": "5765492b-2d11-438b-b9a4-7514ce2ac73e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 773.391253] env[63531]: DEBUG nova.network.neutron [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Successfully updated port: f02115dd-c7b4-46f6-8057-20f0e4985b2e {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 773.408291] env[63531]: DEBUG nova.network.neutron [-] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 773.413039] env[63531]: DEBUG nova.compute.manager [req-fe20945c-f2eb-4528-bc70-187e1f07da58 req-45c4d697-5c71-406a-864b-baa06c630c9e service nova] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Received event network-changed-f02115dd-c7b4-46f6-8057-20f0e4985b2e {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 773.413267] env[63531]: DEBUG nova.compute.manager [req-fe20945c-f2eb-4528-bc70-187e1f07da58 req-45c4d697-5c71-406a-864b-baa06c630c9e service nova] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Refreshing instance network info cache due to event network-changed-f02115dd-c7b4-46f6-8057-20f0e4985b2e. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 773.413547] env[63531]: DEBUG oslo_concurrency.lockutils [req-fe20945c-f2eb-4528-bc70-187e1f07da58 req-45c4d697-5c71-406a-864b-baa06c630c9e service nova] Acquiring lock "refresh_cache-2998fb3d-6677-4fb3-952a-c74d8c6b7b9f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 773.413626] env[63531]: DEBUG oslo_concurrency.lockutils [req-fe20945c-f2eb-4528-bc70-187e1f07da58 req-45c4d697-5c71-406a-864b-baa06c630c9e service nova] Acquired lock "refresh_cache-2998fb3d-6677-4fb3-952a-c74d8c6b7b9f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.413772] env[63531]: DEBUG nova.network.neutron [req-fe20945c-f2eb-4528-bc70-187e1f07da58 req-45c4d697-5c71-406a-864b-baa06c630c9e service nova] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Refreshing network info cache for port f02115dd-c7b4-46f6-8057-20f0e4985b2e {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 773.481457] env[63531]: DEBUG oslo_vmware.api [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Task: {'id': task-1117718, 'name': PowerOffVM_Task, 'duration_secs': 0.214321} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.481979] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 773.481979] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Updating instance '5e9042b3-4dc2-4fa3-a664-c4b49a22e400' progress to 17 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 773.521977] env[63531]: DEBUG oslo_vmware.api [None req-91e9db6d-db1f-484a-9006-e9fab2e8db3e tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Task: {'id': task-1117719, 'name': PowerOffVM_Task, 'duration_secs': 0.215673} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.522989] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-91e9db6d-db1f-484a-9006-e9fab2e8db3e tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] [instance: dccb2335-4220-4570-861b-16ff2d8f5a85] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 773.523178] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-91e9db6d-db1f-484a-9006-e9fab2e8db3e tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] [instance: dccb2335-4220-4570-861b-16ff2d8f5a85] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 773.523420] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a67e8298-67ab-4733-928b-05ba064503ce {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.533396] env[63531]: DEBUG oslo_vmware.api [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]527bf48f-7636-87ee-13f5-add172db02c4, 'name': SearchDatastore_Task, 'duration_secs': 0.009489} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.534106] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-59b092f0-db91-4457-b648-d043ec904d84 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.539388] env[63531]: DEBUG oslo_vmware.api [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 773.539388] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52231d59-8b75-9c15-79e4-670d6754be0c" [ 773.539388] env[63531]: _type = "Task" [ 773.539388] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.546987] env[63531]: DEBUG oslo_vmware.api [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52231d59-8b75-9c15-79e4-670d6754be0c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.591897] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-91e9db6d-db1f-484a-9006-e9fab2e8db3e tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] [instance: dccb2335-4220-4570-861b-16ff2d8f5a85] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 773.592136] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-91e9db6d-db1f-484a-9006-e9fab2e8db3e tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] [instance: dccb2335-4220-4570-861b-16ff2d8f5a85] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 773.592318] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-91e9db6d-db1f-484a-9006-e9fab2e8db3e tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Deleting the datastore file [datastore1] dccb2335-4220-4570-861b-16ff2d8f5a85 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 773.592917] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8078b749-1735-45b4-98a9-c4bfb308724f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.595737] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.344s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 773.596218] env[63531]: DEBUG nova.compute.manager [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 773.600293] env[63531]: DEBUG oslo_concurrency.lockutils [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.843s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 773.602042] env[63531]: INFO nova.compute.claims [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 773.604823] env[63531]: DEBUG oslo_vmware.api [None req-91e9db6d-db1f-484a-9006-e9fab2e8db3e tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Waiting for the task: (returnval){ [ 773.604823] env[63531]: value = "task-1117721" [ 773.604823] env[63531]: _type = "Task" [ 773.604823] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.614155] env[63531]: DEBUG oslo_vmware.api [None req-91e9db6d-db1f-484a-9006-e9fab2e8db3e tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Task: {'id': task-1117721, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.830065] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-aa4c778f-52a8-4043-b6b6-1f33e07173d9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.841708] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3a0d779-0edd-4f55-80ab-7fe95541abdc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.868206] env[63531]: DEBUG nova.compute.manager [req-5db97274-8a7f-486e-a4b2-fa7562bd7370 req-43e20959-febe-45c0-8cc4-653716906ff2 service nova] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Detach interface failed, port_id=4da738f4-b2e9-40c4-9a02-43b6ed658cfa, reason: Instance e1693594-d8bb-4002-b71c-7013efb81d16 could not be found. {{(pid=63531) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 773.868468] env[63531]: DEBUG nova.compute.manager [req-5db97274-8a7f-486e-a4b2-fa7562bd7370 req-43e20959-febe-45c0-8cc4-653716906ff2 service nova] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Received event network-vif-deleted-4c96ebbd-98fa-4d12-a388-b9d270b7aaae {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 773.868639] env[63531]: INFO nova.compute.manager [req-5db97274-8a7f-486e-a4b2-fa7562bd7370 req-43e20959-febe-45c0-8cc4-653716906ff2 service nova] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Neutron deleted interface 4c96ebbd-98fa-4d12-a388-b9d270b7aaae; detaching it from the instance and deleting it from the info cache [ 773.868986] env[63531]: DEBUG nova.network.neutron [req-5db97274-8a7f-486e-a4b2-fa7562bd7370 req-43e20959-febe-45c0-8cc4-653716906ff2 service nova] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Updating instance_info_cache with network_info: [{"id": "5765492b-2d11-438b-b9a4-7514ce2ac73e", "address": "fa:16:3e:f7:f1:ba", "network": {"id": "485000a3-2bad-4c7d-b389-d4200550cc26", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-446289262", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.66", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "8bebde0e281d485ebdd70a1a3b613632", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5765492b-2d", "ovs_interfaceid": "5765492b-2d11-438b-b9a4-7514ce2ac73e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 773.895782] env[63531]: DEBUG oslo_concurrency.lockutils [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Acquiring lock "refresh_cache-2998fb3d-6677-4fb3-952a-c74d8c6b7b9f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 773.910414] env[63531]: INFO nova.compute.manager [-] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Took 2.82 seconds to deallocate network for instance. [ 773.971673] env[63531]: DEBUG nova.network.neutron [req-fe20945c-f2eb-4528-bc70-187e1f07da58 req-45c4d697-5c71-406a-864b-baa06c630c9e service nova] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 773.988935] env[63531]: DEBUG nova.virt.hardware [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 773.991504] env[63531]: DEBUG nova.virt.hardware [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 773.991681] env[63531]: DEBUG nova.virt.hardware [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 773.991880] env[63531]: DEBUG nova.virt.hardware [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 773.992075] env[63531]: DEBUG nova.virt.hardware [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 773.992249] env[63531]: DEBUG nova.virt.hardware [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 773.992461] env[63531]: DEBUG nova.virt.hardware [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 773.992622] env[63531]: DEBUG nova.virt.hardware [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 773.992786] env[63531]: DEBUG nova.virt.hardware [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 773.992949] env[63531]: DEBUG nova.virt.hardware [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 773.993152] env[63531]: DEBUG nova.virt.hardware [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 774.001073] env[63531]: DEBUG nova.network.neutron [req-8e6c56ef-13cd-4f8b-9ccd-559ea27728f9 req-748f5ab8-a2ea-451a-8af1-056b77c0e318 service nova] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Updated VIF entry in instance network info cache for port 262abfb7-7bcf-40e6-9ce3-b502de45b3b4. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 774.001414] env[63531]: DEBUG nova.network.neutron [req-8e6c56ef-13cd-4f8b-9ccd-559ea27728f9 req-748f5ab8-a2ea-451a-8af1-056b77c0e318 service nova] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Updating instance_info_cache with network_info: [{"id": "262abfb7-7bcf-40e6-9ce3-b502de45b3b4", "address": "fa:16:3e:0e:00:c9", "network": {"id": "0667c176-cfdf-4e43-b152-e00bdc224448", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-243715985-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.217", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "308eabafea9c4188a58a0f1c22074d2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8f580e6-1d86-41ee-9ebe-c531cb9299c6", "external-id": "nsx-vlan-transportzone-150", "segmentation_id": 150, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap262abfb7-7b", "ovs_interfaceid": "262abfb7-7bcf-40e6-9ce3-b502de45b3b4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 774.003364] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-74a74c28-d1a1-4390-abd9-5edc7578fb03 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.019512] env[63531]: DEBUG oslo_vmware.api [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Waiting for the task: (returnval){ [ 774.019512] env[63531]: value = "task-1117722" [ 774.019512] env[63531]: _type = "Task" [ 774.019512] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.028513] env[63531]: DEBUG oslo_vmware.api [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Task: {'id': task-1117722, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.063926] env[63531]: DEBUG oslo_vmware.api [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52231d59-8b75-9c15-79e4-670d6754be0c, 'name': SearchDatastore_Task, 'duration_secs': 0.008142} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.064327] env[63531]: DEBUG oslo_concurrency.lockutils [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 774.065216] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 8dec0cac-defc-42f2-ab56-6b3ae60ad858/8dec0cac-defc-42f2-ab56-6b3ae60ad858.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 774.065216] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1b879cc6-7e85-4467-b8ff-f3d2d34ba8c0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.073522] env[63531]: DEBUG oslo_vmware.api [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 774.073522] env[63531]: value = "task-1117723" [ 774.073522] env[63531]: _type = "Task" [ 774.073522] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.082467] env[63531]: DEBUG oslo_vmware.api [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1117723, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.088984] env[63531]: DEBUG nova.network.neutron [req-fe20945c-f2eb-4528-bc70-187e1f07da58 req-45c4d697-5c71-406a-864b-baa06c630c9e service nova] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 774.107010] env[63531]: DEBUG nova.compute.utils [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 774.108535] env[63531]: DEBUG nova.compute.manager [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 774.108710] env[63531]: DEBUG nova.network.neutron [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 774.123108] env[63531]: DEBUG oslo_vmware.api [None req-91e9db6d-db1f-484a-9006-e9fab2e8db3e tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Task: {'id': task-1117721, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.241713} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.123648] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-91e9db6d-db1f-484a-9006-e9fab2e8db3e tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 774.123839] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-91e9db6d-db1f-484a-9006-e9fab2e8db3e tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] [instance: dccb2335-4220-4570-861b-16ff2d8f5a85] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 774.124019] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-91e9db6d-db1f-484a-9006-e9fab2e8db3e tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] [instance: dccb2335-4220-4570-861b-16ff2d8f5a85] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 774.124199] env[63531]: INFO nova.compute.manager [None req-91e9db6d-db1f-484a-9006-e9fab2e8db3e tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] [instance: dccb2335-4220-4570-861b-16ff2d8f5a85] Took 1.16 seconds to destroy the instance on the hypervisor. [ 774.124431] env[63531]: DEBUG oslo.service.loopingcall [None req-91e9db6d-db1f-484a-9006-e9fab2e8db3e tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 774.124617] env[63531]: DEBUG nova.compute.manager [-] [instance: dccb2335-4220-4570-861b-16ff2d8f5a85] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 774.124709] env[63531]: DEBUG nova.network.neutron [-] [instance: dccb2335-4220-4570-861b-16ff2d8f5a85] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 774.212120] env[63531]: DEBUG nova.policy [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd0112e02798343e98ad8a723f458cade', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '401b45e0117342589abdfd44bfcc1a47', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 774.373571] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e5951e8d-a317-4c4a-9070-fce329abc164 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.384729] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ad6159d-fc82-41d2-b63f-d723c7df2984 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.411683] env[63531]: DEBUG nova.compute.manager [req-5db97274-8a7f-486e-a4b2-fa7562bd7370 req-43e20959-febe-45c0-8cc4-653716906ff2 service nova] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Detach interface failed, port_id=4c96ebbd-98fa-4d12-a388-b9d270b7aaae, reason: Instance e1693594-d8bb-4002-b71c-7013efb81d16 could not be found. {{(pid=63531) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 774.412012] env[63531]: DEBUG nova.compute.manager [req-5db97274-8a7f-486e-a4b2-fa7562bd7370 req-43e20959-febe-45c0-8cc4-653716906ff2 service nova] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Received event network-vif-deleted-5765492b-2d11-438b-b9a4-7514ce2ac73e {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 774.412123] env[63531]: INFO nova.compute.manager [req-5db97274-8a7f-486e-a4b2-fa7562bd7370 req-43e20959-febe-45c0-8cc4-653716906ff2 service nova] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Neutron deleted interface 5765492b-2d11-438b-b9a4-7514ce2ac73e; detaching it from the instance and deleting it from the info cache [ 774.412315] env[63531]: DEBUG nova.network.neutron [req-5db97274-8a7f-486e-a4b2-fa7562bd7370 req-43e20959-febe-45c0-8cc4-653716906ff2 service nova] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 774.419962] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7385c1eb-e474-478e-a87d-01f2ddd50615 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 774.514258] env[63531]: DEBUG oslo_concurrency.lockutils [req-8e6c56ef-13cd-4f8b-9ccd-559ea27728f9 req-748f5ab8-a2ea-451a-8af1-056b77c0e318 service nova] Releasing lock "refresh_cache-29dd6de3-2765-408c-acc0-da47e5e0a977" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 774.529467] env[63531]: DEBUG oslo_vmware.api [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Task: {'id': task-1117722, 'name': ReconfigVM_Task, 'duration_secs': 0.221319} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.529765] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Updating instance '5e9042b3-4dc2-4fa3-a664-c4b49a22e400' progress to 33 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 774.583027] env[63531]: DEBUG oslo_vmware.api [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1117723, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.476197} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.583298] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 8dec0cac-defc-42f2-ab56-6b3ae60ad858/8dec0cac-defc-42f2-ab56-6b3ae60ad858.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 774.583535] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 774.583831] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-07b79b55-c70e-480c-ad88-2711cf704d4e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.590935] env[63531]: DEBUG oslo_vmware.api [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 774.590935] env[63531]: value = "task-1117724" [ 774.590935] env[63531]: _type = "Task" [ 774.590935] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.593573] env[63531]: DEBUG oslo_concurrency.lockutils [req-fe20945c-f2eb-4528-bc70-187e1f07da58 req-45c4d697-5c71-406a-864b-baa06c630c9e service nova] Releasing lock "refresh_cache-2998fb3d-6677-4fb3-952a-c74d8c6b7b9f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 774.593917] env[63531]: DEBUG oslo_concurrency.lockutils [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Acquired lock "refresh_cache-2998fb3d-6677-4fb3-952a-c74d8c6b7b9f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 774.594082] env[63531]: DEBUG nova.network.neutron [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 774.601962] env[63531]: DEBUG oslo_vmware.api [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1117724, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.615391] env[63531]: DEBUG nova.compute.manager [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 774.740034] env[63531]: DEBUG nova.network.neutron [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Successfully created port: 96e03393-975e-4587-b22c-59fe3ba65718 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 774.901858] env[63531]: DEBUG nova.network.neutron [-] [instance: dccb2335-4220-4570-861b-16ff2d8f5a85] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 774.918469] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-801fe976-49fa-41be-a91c-e7557a1963ad {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.930730] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8986c6cb-64e0-4d78-a8e8-f4bb6fb327bd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.958955] env[63531]: DEBUG nova.compute.manager [req-5db97274-8a7f-486e-a4b2-fa7562bd7370 req-43e20959-febe-45c0-8cc4-653716906ff2 service nova] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Detach interface failed, port_id=5765492b-2d11-438b-b9a4-7514ce2ac73e, reason: Instance e1693594-d8bb-4002-b71c-7013efb81d16 could not be found. {{(pid=63531) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 775.009716] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f21a58ce-c497-46ab-ad7c-561c48b34879 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.017221] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae2a4fa7-69ff-400c-a606-f4d942327ea6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.048520] env[63531]: DEBUG nova.virt.hardware [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:17:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='86bdb6e9-50cf-4ad6-bad7-f8cb5dc21d1b',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-672645615',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 775.048753] env[63531]: DEBUG nova.virt.hardware [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 775.048904] env[63531]: DEBUG nova.virt.hardware [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 775.049096] env[63531]: DEBUG nova.virt.hardware [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 775.049241] env[63531]: DEBUG nova.virt.hardware [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 775.049385] env[63531]: DEBUG nova.virt.hardware [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 775.049584] env[63531]: DEBUG nova.virt.hardware [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 775.049738] env[63531]: DEBUG nova.virt.hardware [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 775.049899] env[63531]: DEBUG nova.virt.hardware [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 775.050070] env[63531]: DEBUG nova.virt.hardware [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 775.050244] env[63531]: DEBUG nova.virt.hardware [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 775.055389] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Reconfiguring VM instance instance-00000027 to detach disk 2000 {{(pid=63531) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 775.056085] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3346934e-a0db-4d2d-abae-68ca5f609b95 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.068370] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ddcd2db-e5be-4ed9-b6b6-98f198b78aee {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.079982] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78dd88ab-9dec-4f46-873d-239bd8f3f3db {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.086016] env[63531]: DEBUG oslo_vmware.api [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Waiting for the task: (returnval){ [ 775.086016] env[63531]: value = "task-1117725" [ 775.086016] env[63531]: _type = "Task" [ 775.086016] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.104639] env[63531]: DEBUG nova.compute.provider_tree [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Updating inventory in ProviderTree for provider 9996a314-a62e-4282-8252-f387e7e3fabb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 775.117062] env[63531]: DEBUG oslo_vmware.api [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Task: {'id': task-1117725, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.125823] env[63531]: DEBUG oslo_vmware.api [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1117724, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066627} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.126378] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 775.127245] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d14fbb2d-f8b8-4980-9b93-f6204fa1f504 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.150262] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Reconfiguring VM instance instance-00000031 to attach disk [datastore1] 8dec0cac-defc-42f2-ab56-6b3ae60ad858/8dec0cac-defc-42f2-ab56-6b3ae60ad858.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 775.151245] env[63531]: DEBUG nova.network.neutron [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 775.153182] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-69e74f19-cc5e-49f5-8806-69aeb36f09a7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.174096] env[63531]: DEBUG oslo_vmware.api [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 775.174096] env[63531]: value = "task-1117726" [ 775.174096] env[63531]: _type = "Task" [ 775.174096] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.183509] env[63531]: DEBUG oslo_vmware.api [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1117726, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.319536] env[63531]: DEBUG nova.network.neutron [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Updating instance_info_cache with network_info: [{"id": "f02115dd-c7b4-46f6-8057-20f0e4985b2e", "address": "fa:16:3e:4e:ff:37", "network": {"id": "5439232a-ec6d-4184-a9f1-f7be5c1526bc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.232", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "54fbef8a6ee0488f820706d2036e9c99", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f267bcdd-0daa-4337-9709-5fc060c267d8", "external-id": "nsx-vlan-transportzone-308", "segmentation_id": 308, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf02115dd-c7", "ovs_interfaceid": "f02115dd-c7b4-46f6-8057-20f0e4985b2e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 775.404024] env[63531]: INFO nova.compute.manager [-] [instance: dccb2335-4220-4570-861b-16ff2d8f5a85] Took 1.28 seconds to deallocate network for instance. [ 775.441900] env[63531]: DEBUG nova.compute.manager [req-a3131941-299b-4ab3-b84b-d8d400811a56 req-c239f177-f750-4407-9a46-b4635216ab62 service nova] [instance: dccb2335-4220-4570-861b-16ff2d8f5a85] Received event network-vif-deleted-d69ccbf2-4502-4643-92b4-6fad825e76b4 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 775.595844] env[63531]: DEBUG oslo_vmware.api [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Task: {'id': task-1117725, 'name': ReconfigVM_Task, 'duration_secs': 0.312228} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.596148] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Reconfigured VM instance instance-00000027 to detach disk 2000 {{(pid=63531) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 775.596921] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cace32ea-72c4-42db-b140-4fad6fc7e8d6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.620663] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Reconfiguring VM instance instance-00000027 to attach disk [datastore1] 5e9042b3-4dc2-4fa3-a664-c4b49a22e400/5e9042b3-4dc2-4fa3-a664-c4b49a22e400.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 775.623214] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b9819e87-5a4d-4bde-8af4-2770e577d513 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.636775] env[63531]: DEBUG nova.compute.manager [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 775.644343] env[63531]: DEBUG oslo_vmware.api [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Waiting for the task: (returnval){ [ 775.644343] env[63531]: value = "task-1117727" [ 775.644343] env[63531]: _type = "Task" [ 775.644343] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.654545] env[63531]: DEBUG oslo_vmware.api [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Task: {'id': task-1117727, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.662821] env[63531]: DEBUG nova.virt.hardware [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 775.663118] env[63531]: DEBUG nova.virt.hardware [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 775.663318] env[63531]: DEBUG nova.virt.hardware [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 775.663536] env[63531]: DEBUG nova.virt.hardware [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 775.663703] env[63531]: DEBUG nova.virt.hardware [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 775.663857] env[63531]: DEBUG nova.virt.hardware [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 775.664088] env[63531]: DEBUG nova.virt.hardware [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 775.664256] env[63531]: DEBUG nova.virt.hardware [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 775.664422] env[63531]: DEBUG nova.virt.hardware [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 775.664581] env[63531]: DEBUG nova.virt.hardware [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 775.664752] env[63531]: DEBUG nova.virt.hardware [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 775.665626] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdf3b570-2ac2-488e-b1c0-df368f1f4d89 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.668960] env[63531]: DEBUG nova.scheduler.client.report [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Updated inventory for provider 9996a314-a62e-4282-8252-f387e7e3fabb with generation 73 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 775.669229] env[63531]: DEBUG nova.compute.provider_tree [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Updating resource provider 9996a314-a62e-4282-8252-f387e7e3fabb generation from 73 to 74 during operation: update_inventory {{(pid=63531) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 775.669389] env[63531]: DEBUG nova.compute.provider_tree [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Updating inventory in ProviderTree for provider 9996a314-a62e-4282-8252-f387e7e3fabb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 775.681008] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d2d2429-bd9a-4e76-ba29-a9e18c58f0c1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.700019] env[63531]: DEBUG oslo_vmware.api [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1117726, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.822993] env[63531]: DEBUG oslo_concurrency.lockutils [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Releasing lock "refresh_cache-2998fb3d-6677-4fb3-952a-c74d8c6b7b9f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 775.823441] env[63531]: DEBUG nova.compute.manager [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Instance network_info: |[{"id": "f02115dd-c7b4-46f6-8057-20f0e4985b2e", "address": "fa:16:3e:4e:ff:37", "network": {"id": "5439232a-ec6d-4184-a9f1-f7be5c1526bc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.232", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "54fbef8a6ee0488f820706d2036e9c99", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f267bcdd-0daa-4337-9709-5fc060c267d8", "external-id": "nsx-vlan-transportzone-308", "segmentation_id": 308, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf02115dd-c7", "ovs_interfaceid": "f02115dd-c7b4-46f6-8057-20f0e4985b2e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 775.824340] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4e:ff:37', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f267bcdd-0daa-4337-9709-5fc060c267d8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f02115dd-c7b4-46f6-8057-20f0e4985b2e', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 775.832225] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Creating folder: Project (401b45e0117342589abdfd44bfcc1a47). Parent ref: group-v244585. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 775.832526] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-beb6a41e-ac69-4bcb-9d2a-aea8c9210899 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.844043] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Created folder: Project (401b45e0117342589abdfd44bfcc1a47) in parent group-v244585. [ 775.844043] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Creating folder: Instances. Parent ref: group-v244652. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 775.844043] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-295fe410-6a0d-4a38-a482-b53792ecb775 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.853306] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Created folder: Instances in parent group-v244652. [ 775.853526] env[63531]: DEBUG oslo.service.loopingcall [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 775.853708] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 775.853906] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f324cf95-a7fe-43c7-b29b-ba415f68ed44 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.872827] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 775.872827] env[63531]: value = "task-1117730" [ 775.872827] env[63531]: _type = "Task" [ 775.872827] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.880324] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117730, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.911635] env[63531]: DEBUG oslo_concurrency.lockutils [None req-91e9db6d-db1f-484a-9006-e9fab2e8db3e tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 776.153976] env[63531]: DEBUG oslo_vmware.api [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Task: {'id': task-1117727, 'name': ReconfigVM_Task, 'duration_secs': 0.288016} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.158631] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Reconfigured VM instance instance-00000027 to attach disk [datastore1] 5e9042b3-4dc2-4fa3-a664-c4b49a22e400/5e9042b3-4dc2-4fa3-a664-c4b49a22e400.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 776.158631] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Updating instance '5e9042b3-4dc2-4fa3-a664-c4b49a22e400' progress to 50 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 776.177806] env[63531]: DEBUG oslo_concurrency.lockutils [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.577s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.178355] env[63531]: DEBUG nova.compute.manager [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 776.185377] env[63531]: DEBUG oslo_concurrency.lockutils [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.065s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 776.187267] env[63531]: INFO nova.compute.claims [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 776.196709] env[63531]: DEBUG oslo_vmware.api [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1117726, 'name': ReconfigVM_Task, 'duration_secs': 0.933674} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.197090] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Reconfigured VM instance instance-00000031 to attach disk [datastore1] 8dec0cac-defc-42f2-ab56-6b3ae60ad858/8dec0cac-defc-42f2-ab56-6b3ae60ad858.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 776.197713] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cf15b805-14af-454d-b6ed-a43b70c4a857 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.204049] env[63531]: DEBUG oslo_vmware.api [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 776.204049] env[63531]: value = "task-1117731" [ 776.204049] env[63531]: _type = "Task" [ 776.204049] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.212406] env[63531]: DEBUG oslo_vmware.api [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1117731, 'name': Rename_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.255942] env[63531]: DEBUG nova.compute.manager [req-d620104f-e8b1-4258-ab22-dc69f62f64eb req-865d6880-2fb6-4dae-b34d-299a9eca44b1 service nova] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Received event network-vif-plugged-96e03393-975e-4587-b22c-59fe3ba65718 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 776.255942] env[63531]: DEBUG oslo_concurrency.lockutils [req-d620104f-e8b1-4258-ab22-dc69f62f64eb req-865d6880-2fb6-4dae-b34d-299a9eca44b1 service nova] Acquiring lock "64fb5831-0789-482b-bb98-67b29868c4c9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 776.255942] env[63531]: DEBUG oslo_concurrency.lockutils [req-d620104f-e8b1-4258-ab22-dc69f62f64eb req-865d6880-2fb6-4dae-b34d-299a9eca44b1 service nova] Lock "64fb5831-0789-482b-bb98-67b29868c4c9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 776.256425] env[63531]: DEBUG oslo_concurrency.lockutils [req-d620104f-e8b1-4258-ab22-dc69f62f64eb req-865d6880-2fb6-4dae-b34d-299a9eca44b1 service nova] Lock "64fb5831-0789-482b-bb98-67b29868c4c9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.256810] env[63531]: DEBUG nova.compute.manager [req-d620104f-e8b1-4258-ab22-dc69f62f64eb req-865d6880-2fb6-4dae-b34d-299a9eca44b1 service nova] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] No waiting events found dispatching network-vif-plugged-96e03393-975e-4587-b22c-59fe3ba65718 {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 776.257255] env[63531]: WARNING nova.compute.manager [req-d620104f-e8b1-4258-ab22-dc69f62f64eb req-865d6880-2fb6-4dae-b34d-299a9eca44b1 service nova] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Received unexpected event network-vif-plugged-96e03393-975e-4587-b22c-59fe3ba65718 for instance with vm_state building and task_state spawning. [ 776.367242] env[63531]: DEBUG nova.network.neutron [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Successfully updated port: 96e03393-975e-4587-b22c-59fe3ba65718 {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 776.384912] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117730, 'name': CreateVM_Task, 'duration_secs': 0.286961} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.385134] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 776.385819] env[63531]: DEBUG oslo_concurrency.lockutils [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.385981] env[63531]: DEBUG oslo_concurrency.lockutils [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.386328] env[63531]: DEBUG oslo_concurrency.lockutils [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 776.386840] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c3d66db3-684e-4676-8582-9c77bbb063f3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.391995] env[63531]: DEBUG oslo_vmware.api [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Waiting for the task: (returnval){ [ 776.391995] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52bb7a45-7b98-a84a-c95d-e00f428ba80b" [ 776.391995] env[63531]: _type = "Task" [ 776.391995] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.399930] env[63531]: DEBUG oslo_vmware.api [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52bb7a45-7b98-a84a-c95d-e00f428ba80b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.662627] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daa3fc07-f2f6-4c56-80bd-74ae09b30501 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.682532] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43f83b7e-cba5-41f7-8208-e173a84302a1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.703080] env[63531]: DEBUG nova.compute.utils [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 776.704209] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Updating instance '5e9042b3-4dc2-4fa3-a664-c4b49a22e400' progress to 67 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 776.707579] env[63531]: DEBUG nova.compute.manager [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 776.707747] env[63531]: DEBUG nova.network.neutron [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 776.720017] env[63531]: DEBUG oslo_vmware.api [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1117731, 'name': Rename_Task, 'duration_secs': 0.318734} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.720017] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 776.720017] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cd0e0343-4a3b-4e01-ad0f-4a6293a0fc33 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.726251] env[63531]: DEBUG oslo_vmware.api [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 776.726251] env[63531]: value = "task-1117732" [ 776.726251] env[63531]: _type = "Task" [ 776.726251] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.734471] env[63531]: DEBUG oslo_vmware.api [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1117732, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.808343] env[63531]: DEBUG nova.policy [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c75075d6c01740abbfcd716415da183f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '68eaa455a44b4d96936fe8273629279a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 776.872589] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Acquiring lock "refresh_cache-64fb5831-0789-482b-bb98-67b29868c4c9" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.872589] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Acquired lock "refresh_cache-64fb5831-0789-482b-bb98-67b29868c4c9" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.872589] env[63531]: DEBUG nova.network.neutron [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 776.903646] env[63531]: DEBUG oslo_vmware.api [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52bb7a45-7b98-a84a-c95d-e00f428ba80b, 'name': SearchDatastore_Task, 'duration_secs': 0.009598} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.904015] env[63531]: DEBUG oslo_concurrency.lockutils [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 776.904344] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 776.904841] env[63531]: DEBUG oslo_concurrency.lockutils [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.904841] env[63531]: DEBUG oslo_concurrency.lockutils [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.905051] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 776.905403] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a7d47479-e549-4dc6-ad4c-c0b340868ec1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.915747] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 776.916105] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 776.919934] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-25f2b42c-de2b-4cc7-b54e-60dd2c0c06f0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.930683] env[63531]: DEBUG oslo_vmware.api [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Waiting for the task: (returnval){ [ 776.930683] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52788ad8-d761-816a-0fab-a00757f18ade" [ 776.930683] env[63531]: _type = "Task" [ 776.930683] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.939240] env[63531]: DEBUG oslo_vmware.api [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52788ad8-d761-816a-0fab-a00757f18ade, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.201734] env[63531]: DEBUG nova.network.neutron [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Successfully created port: b4141cf4-dab5-4033-8cc5-06966340323b {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 777.210655] env[63531]: DEBUG nova.compute.manager [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 777.236927] env[63531]: DEBUG oslo_vmware.api [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1117732, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.255833] env[63531]: DEBUG nova.network.neutron [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Port f178c862-e753-40d4-b7cd-3eabe7eda39e binding to destination host cpu-1 is already ACTIVE {{(pid=63531) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 777.413919] env[63531]: DEBUG nova.network.neutron [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 777.441109] env[63531]: DEBUG oslo_vmware.api [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52788ad8-d761-816a-0fab-a00757f18ade, 'name': SearchDatastore_Task, 'duration_secs': 0.010566} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.446255] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5b199e61-1d93-410a-aec5-2223a3baaaaf {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.456066] env[63531]: DEBUG oslo_vmware.api [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Waiting for the task: (returnval){ [ 777.456066] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5241d6f2-9430-920a-835b-84e664f59f8f" [ 777.456066] env[63531]: _type = "Task" [ 777.456066] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.465122] env[63531]: DEBUG oslo_vmware.api [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5241d6f2-9430-920a-835b-84e664f59f8f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.607956] env[63531]: DEBUG nova.network.neutron [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Updating instance_info_cache with network_info: [{"id": "96e03393-975e-4587-b22c-59fe3ba65718", "address": "fa:16:3e:36:f7:d3", "network": {"id": "5439232a-ec6d-4184-a9f1-f7be5c1526bc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.57", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "54fbef8a6ee0488f820706d2036e9c99", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f267bcdd-0daa-4337-9709-5fc060c267d8", "external-id": "nsx-vlan-transportzone-308", "segmentation_id": 308, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap96e03393-97", "ovs_interfaceid": "96e03393-975e-4587-b22c-59fe3ba65718", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.628238] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c312223-d839-407c-b768-5d28927adf05 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.637675] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48afad15-02a4-4fac-8811-f7376228b324 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.668978] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4df711a-75b7-4f01-b6a6-ee3d339904c9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.677370] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7a2b6e5-fb1a-4e88-a903-fd926c735bfd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.696034] env[63531]: DEBUG nova.compute.provider_tree [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 777.736358] env[63531]: DEBUG oslo_vmware.api [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1117732, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.966675] env[63531]: DEBUG oslo_vmware.api [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5241d6f2-9430-920a-835b-84e664f59f8f, 'name': SearchDatastore_Task, 'duration_secs': 0.009612} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.966964] env[63531]: DEBUG oslo_concurrency.lockutils [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 777.967268] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f/2998fb3d-6677-4fb3-952a-c74d8c6b7b9f.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 777.967527] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-970b071b-8dfa-46e0-ac9f-f46025750174 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.974882] env[63531]: DEBUG oslo_vmware.api [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Waiting for the task: (returnval){ [ 777.974882] env[63531]: value = "task-1117733" [ 777.974882] env[63531]: _type = "Task" [ 777.974882] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.982829] env[63531]: DEBUG oslo_vmware.api [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117733, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.110681] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Releasing lock "refresh_cache-64fb5831-0789-482b-bb98-67b29868c4c9" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.111093] env[63531]: DEBUG nova.compute.manager [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Instance network_info: |[{"id": "96e03393-975e-4587-b22c-59fe3ba65718", "address": "fa:16:3e:36:f7:d3", "network": {"id": "5439232a-ec6d-4184-a9f1-f7be5c1526bc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.57", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "54fbef8a6ee0488f820706d2036e9c99", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f267bcdd-0daa-4337-9709-5fc060c267d8", "external-id": "nsx-vlan-transportzone-308", "segmentation_id": 308, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap96e03393-97", "ovs_interfaceid": "96e03393-975e-4587-b22c-59fe3ba65718", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 778.111561] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:36:f7:d3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f267bcdd-0daa-4337-9709-5fc060c267d8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '96e03393-975e-4587-b22c-59fe3ba65718', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 778.119226] env[63531]: DEBUG oslo.service.loopingcall [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 778.119447] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 778.119685] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-53ef3725-160a-4b26-af6c-962c2c8e6167 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.140324] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 778.140324] env[63531]: value = "task-1117734" [ 778.140324] env[63531]: _type = "Task" [ 778.140324] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.150052] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117734, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.200194] env[63531]: DEBUG nova.scheduler.client.report [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 778.222925] env[63531]: DEBUG nova.compute.manager [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 778.242580] env[63531]: DEBUG oslo_vmware.api [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1117732, 'name': PowerOnVM_Task, 'duration_secs': 1.213917} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.243018] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 778.243434] env[63531]: INFO nova.compute.manager [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Took 10.50 seconds to spawn the instance on the hypervisor. [ 778.243807] env[63531]: DEBUG nova.compute.manager [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 778.247837] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ba10e7d-78cf-4d3b-9387-a058c5b98d7c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.262148] env[63531]: DEBUG nova.virt.hardware [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 778.262423] env[63531]: DEBUG nova.virt.hardware [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 778.262584] env[63531]: DEBUG nova.virt.hardware [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 778.262760] env[63531]: DEBUG nova.virt.hardware [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 778.262904] env[63531]: DEBUG nova.virt.hardware [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 778.263062] env[63531]: DEBUG nova.virt.hardware [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 778.263374] env[63531]: DEBUG nova.virt.hardware [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 778.263456] env[63531]: DEBUG nova.virt.hardware [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 778.263591] env[63531]: DEBUG nova.virt.hardware [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 778.263747] env[63531]: DEBUG nova.virt.hardware [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 778.263924] env[63531]: DEBUG nova.virt.hardware [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 778.264823] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faabcaa4-8535-4dfe-9b52-cebd207265fb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.286034] env[63531]: DEBUG oslo_concurrency.lockutils [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Acquiring lock "5e9042b3-4dc2-4fa3-a664-c4b49a22e400-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 778.286259] env[63531]: DEBUG oslo_concurrency.lockutils [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Lock "5e9042b3-4dc2-4fa3-a664-c4b49a22e400-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 778.286332] env[63531]: DEBUG oslo_concurrency.lockutils [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Lock "5e9042b3-4dc2-4fa3-a664-c4b49a22e400-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 778.289984] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18d84d8d-893e-429e-a6e7-4186c86c230a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.297587] env[63531]: DEBUG nova.compute.manager [req-649fc902-316f-41d3-ae74-32c50242e55c req-cf038adf-c4f8-40f1-bd1e-d905be52dcc9 service nova] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Received event network-changed-96e03393-975e-4587-b22c-59fe3ba65718 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 778.297772] env[63531]: DEBUG nova.compute.manager [req-649fc902-316f-41d3-ae74-32c50242e55c req-cf038adf-c4f8-40f1-bd1e-d905be52dcc9 service nova] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Refreshing instance network info cache due to event network-changed-96e03393-975e-4587-b22c-59fe3ba65718. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 778.298029] env[63531]: DEBUG oslo_concurrency.lockutils [req-649fc902-316f-41d3-ae74-32c50242e55c req-cf038adf-c4f8-40f1-bd1e-d905be52dcc9 service nova] Acquiring lock "refresh_cache-64fb5831-0789-482b-bb98-67b29868c4c9" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 778.298178] env[63531]: DEBUG oslo_concurrency.lockutils [req-649fc902-316f-41d3-ae74-32c50242e55c req-cf038adf-c4f8-40f1-bd1e-d905be52dcc9 service nova] Acquired lock "refresh_cache-64fb5831-0789-482b-bb98-67b29868c4c9" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.298341] env[63531]: DEBUG nova.network.neutron [req-649fc902-316f-41d3-ae74-32c50242e55c req-cf038adf-c4f8-40f1-bd1e-d905be52dcc9 service nova] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Refreshing network info cache for port 96e03393-975e-4587-b22c-59fe3ba65718 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 778.485997] env[63531]: DEBUG oslo_vmware.api [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117733, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.492853} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.485997] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f/2998fb3d-6677-4fb3-952a-c74d8c6b7b9f.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 778.485997] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 778.486238] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e713bfc8-319d-4102-925c-41c1b41c3404 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.492381] env[63531]: DEBUG oslo_vmware.api [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Waiting for the task: (returnval){ [ 778.492381] env[63531]: value = "task-1117735" [ 778.492381] env[63531]: _type = "Task" [ 778.492381] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.502300] env[63531]: DEBUG oslo_vmware.api [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117735, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.559183] env[63531]: INFO nova.compute.manager [None req-637af179-656f-4b75-bdcc-db626dfdd5bd tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Rebuilding instance [ 778.603667] env[63531]: DEBUG nova.compute.manager [None req-637af179-656f-4b75-bdcc-db626dfdd5bd tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 778.604577] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1081a603-b889-47b1-a333-f862e045c756 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.651860] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117734, 'name': CreateVM_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.707162] env[63531]: DEBUG oslo_concurrency.lockutils [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.522s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 778.707640] env[63531]: DEBUG nova.compute.manager [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 778.710386] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.576s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 778.711796] env[63531]: INFO nova.compute.claims [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 778.774652] env[63531]: INFO nova.compute.manager [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Took 39.40 seconds to build instance. [ 779.005019] env[63531]: DEBUG oslo_vmware.api [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117735, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.092587} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.005019] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 779.005019] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eec895ee-5613-4f89-aef2-77a2ea867ad5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.008251] env[63531]: DEBUG nova.network.neutron [req-649fc902-316f-41d3-ae74-32c50242e55c req-cf038adf-c4f8-40f1-bd1e-d905be52dcc9 service nova] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Updated VIF entry in instance network info cache for port 96e03393-975e-4587-b22c-59fe3ba65718. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 779.008847] env[63531]: DEBUG nova.network.neutron [req-649fc902-316f-41d3-ae74-32c50242e55c req-cf038adf-c4f8-40f1-bd1e-d905be52dcc9 service nova] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Updating instance_info_cache with network_info: [{"id": "96e03393-975e-4587-b22c-59fe3ba65718", "address": "fa:16:3e:36:f7:d3", "network": {"id": "5439232a-ec6d-4184-a9f1-f7be5c1526bc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.57", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "54fbef8a6ee0488f820706d2036e9c99", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f267bcdd-0daa-4337-9709-5fc060c267d8", "external-id": "nsx-vlan-transportzone-308", "segmentation_id": 308, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap96e03393-97", "ovs_interfaceid": "96e03393-975e-4587-b22c-59fe3ba65718", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.031078] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Reconfiguring VM instance instance-00000032 to attach disk [datastore1] 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f/2998fb3d-6677-4fb3-952a-c74d8c6b7b9f.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 779.031832] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fe2c0695-3514-4e95-ba75-f756091b44b3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.054272] env[63531]: DEBUG oslo_vmware.api [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Waiting for the task: (returnval){ [ 779.054272] env[63531]: value = "task-1117736" [ 779.054272] env[63531]: _type = "Task" [ 779.054272] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.062490] env[63531]: DEBUG oslo_vmware.api [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117736, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.099284] env[63531]: DEBUG nova.network.neutron [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Successfully updated port: b4141cf4-dab5-4033-8cc5-06966340323b {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 779.115772] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-637af179-656f-4b75-bdcc-db626dfdd5bd tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 779.116084] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-78a493d4-0deb-444b-8d50-319ff3e17f4f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.124454] env[63531]: DEBUG oslo_vmware.api [None req-637af179-656f-4b75-bdcc-db626dfdd5bd tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Waiting for the task: (returnval){ [ 779.124454] env[63531]: value = "task-1117737" [ 779.124454] env[63531]: _type = "Task" [ 779.124454] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.134428] env[63531]: DEBUG oslo_vmware.api [None req-637af179-656f-4b75-bdcc-db626dfdd5bd tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Task: {'id': task-1117737, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.150381] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117734, 'name': CreateVM_Task, 'duration_secs': 0.566799} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.150560] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 779.151228] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.151418] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.151747] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 779.152032] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7a21972a-3b87-4267-9ddc-4b9bf801ca8e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.157020] env[63531]: DEBUG oslo_vmware.api [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Waiting for the task: (returnval){ [ 779.157020] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5292b7e2-88d6-cd29-9d19-90884e0d93b9" [ 779.157020] env[63531]: _type = "Task" [ 779.157020] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.164374] env[63531]: DEBUG oslo_vmware.api [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5292b7e2-88d6-cd29-9d19-90884e0d93b9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.219300] env[63531]: DEBUG nova.compute.utils [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 779.220115] env[63531]: DEBUG nova.compute.manager [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 779.220305] env[63531]: DEBUG nova.network.neutron [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 779.265442] env[63531]: DEBUG nova.policy [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '30c2b29064b44c48ad7e76d445c65317', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '04a4c814245346a9b73253a25a822fef', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 779.277746] env[63531]: DEBUG oslo_concurrency.lockutils [None req-201ee6cd-0fc4-4efd-8246-7ade220681c8 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "8dec0cac-defc-42f2-ab56-6b3ae60ad858" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 125.470s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.344942] env[63531]: DEBUG oslo_concurrency.lockutils [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Acquiring lock "refresh_cache-5e9042b3-4dc2-4fa3-a664-c4b49a22e400" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.345181] env[63531]: DEBUG oslo_concurrency.lockutils [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Acquired lock "refresh_cache-5e9042b3-4dc2-4fa3-a664-c4b49a22e400" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.345361] env[63531]: DEBUG nova.network.neutron [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 779.513061] env[63531]: DEBUG oslo_concurrency.lockutils [req-649fc902-316f-41d3-ae74-32c50242e55c req-cf038adf-c4f8-40f1-bd1e-d905be52dcc9 service nova] Releasing lock "refresh_cache-64fb5831-0789-482b-bb98-67b29868c4c9" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.564822] env[63531]: DEBUG oslo_vmware.api [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117736, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.602101] env[63531]: DEBUG oslo_concurrency.lockutils [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Acquiring lock "refresh_cache-0197e03d-6c36-4e73-9472-8e3319eb89ce" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.602277] env[63531]: DEBUG oslo_concurrency.lockutils [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Acquired lock "refresh_cache-0197e03d-6c36-4e73-9472-8e3319eb89ce" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.602432] env[63531]: DEBUG nova.network.neutron [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 779.634311] env[63531]: DEBUG oslo_vmware.api [None req-637af179-656f-4b75-bdcc-db626dfdd5bd tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Task: {'id': task-1117737, 'name': PowerOffVM_Task, 'duration_secs': 0.20364} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.634585] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-637af179-656f-4b75-bdcc-db626dfdd5bd tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 779.635285] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-637af179-656f-4b75-bdcc-db626dfdd5bd tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 779.635529] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5400d8df-1563-4d0c-a965-b83871b7d30f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.642959] env[63531]: DEBUG oslo_vmware.api [None req-637af179-656f-4b75-bdcc-db626dfdd5bd tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Waiting for the task: (returnval){ [ 779.642959] env[63531]: value = "task-1117738" [ 779.642959] env[63531]: _type = "Task" [ 779.642959] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.650487] env[63531]: DEBUG oslo_vmware.api [None req-637af179-656f-4b75-bdcc-db626dfdd5bd tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Task: {'id': task-1117738, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.668241] env[63531]: DEBUG oslo_vmware.api [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5292b7e2-88d6-cd29-9d19-90884e0d93b9, 'name': SearchDatastore_Task, 'duration_secs': 0.009773} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.668721] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.669111] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 779.671104] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.671104] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.671104] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 779.671104] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0109b2af-192a-491c-a353-740aa7be176f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.681205] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 779.681205] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 779.681205] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-049acb1c-7357-4fcf-b886-3a027ce3c3c3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.685789] env[63531]: DEBUG oslo_vmware.api [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Waiting for the task: (returnval){ [ 779.685789] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52ccacd8-0341-c522-7f86-1c0737f4fa20" [ 779.685789] env[63531]: _type = "Task" [ 779.685789] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.694884] env[63531]: DEBUG oslo_vmware.api [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52ccacd8-0341-c522-7f86-1c0737f4fa20, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.724186] env[63531]: DEBUG nova.compute.manager [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 779.781513] env[63531]: DEBUG nova.compute.manager [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] [instance: eb6c84d7-e0e6-4999-9018-889b99dc67ad] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 779.822558] env[63531]: DEBUG nova.network.neutron [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Successfully created port: d4b7599b-ca73-4bab-91d4-33a977545afc {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 780.064495] env[63531]: DEBUG oslo_vmware.api [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117736, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.148920] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5556820-8166-4574-bddd-d4e46f0475b7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.159950] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-637af179-656f-4b75-bdcc-db626dfdd5bd tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] VM already powered off {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 780.160199] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-637af179-656f-4b75-bdcc-db626dfdd5bd tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Volume detach. Driver type: vmdk {{(pid=63531) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 780.160435] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-637af179-656f-4b75-bdcc-db626dfdd5bd tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244601', 'volume_id': '2a24c282-f681-40bf-a19c-5c711d342687', 'name': 'volume-2a24c282-f681-40bf-a19c-5c711d342687', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '410ddf7a-e45b-4769-a3db-5363ce2096a7', 'attached_at': '', 'detached_at': '', 'volume_id': '2a24c282-f681-40bf-a19c-5c711d342687', 'serial': '2a24c282-f681-40bf-a19c-5c711d342687'} {{(pid=63531) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 780.161184] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f62f207-e5f7-4881-93d0-2072cb864fe4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.164386] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cedae158-a6c1-4d03-a439-0ba2cd6308ac {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.211686] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c89acda-9155-4157-b535-23b81ddb05ae {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.217739] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d1c49ac-007a-48da-b9d8-c5a74c42f7e3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.229291] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f652b06-4c58-47fe-9809-5abdef28e0a7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.231631] env[63531]: DEBUG oslo_vmware.api [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52ccacd8-0341-c522-7f86-1c0737f4fa20, 'name': SearchDatastore_Task, 'duration_secs': 0.008481} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.233162] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40bb3014-ac2a-4b50-8035-d1f91a7b4177 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.247325] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2e59e9cf-b73b-4c35-9e30-976bc0440a60 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.264576] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f74bebb4-e05e-4a1f-9d8b-5fbccf2123ae {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.277032] env[63531]: DEBUG nova.compute.provider_tree [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 780.280140] env[63531]: DEBUG oslo_vmware.api [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Waiting for the task: (returnval){ [ 780.280140] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5226c040-bb6e-a5f5-966b-d38ba6e0be4d" [ 780.280140] env[63531]: _type = "Task" [ 780.280140] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.297991] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-637af179-656f-4b75-bdcc-db626dfdd5bd tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] The volume has not been displaced from its original location: [datastore1] volume-2a24c282-f681-40bf-a19c-5c711d342687/volume-2a24c282-f681-40bf-a19c-5c711d342687.vmdk. No consolidation needed. {{(pid=63531) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 780.304461] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-637af179-656f-4b75-bdcc-db626dfdd5bd tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Reconfiguring VM instance instance-00000029 to detach disk 2000 {{(pid=63531) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 780.309314] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9ff0a5c7-e8b9-484d-a2a7-2fe73dd2ca49 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.327461] env[63531]: DEBUG nova.network.neutron [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 780.329830] env[63531]: DEBUG oslo_vmware.api [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5226c040-bb6e-a5f5-966b-d38ba6e0be4d, 'name': SearchDatastore_Task, 'duration_secs': 0.009387} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.330406] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 780.330656] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 64fb5831-0789-482b-bb98-67b29868c4c9/64fb5831-0789-482b-bb98-67b29868c4c9.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 780.331153] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-32c941db-4091-49a2-b5dd-36a2b0af50ef {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.335524] env[63531]: DEBUG oslo_vmware.api [None req-637af179-656f-4b75-bdcc-db626dfdd5bd tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Waiting for the task: (returnval){ [ 780.335524] env[63531]: value = "task-1117739" [ 780.335524] env[63531]: _type = "Task" [ 780.335524] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.336428] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 780.340815] env[63531]: DEBUG oslo_vmware.api [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Waiting for the task: (returnval){ [ 780.340815] env[63531]: value = "task-1117740" [ 780.340815] env[63531]: _type = "Task" [ 780.340815] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.346947] env[63531]: DEBUG oslo_vmware.api [None req-637af179-656f-4b75-bdcc-db626dfdd5bd tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Task: {'id': task-1117739, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.351430] env[63531]: DEBUG oslo_vmware.api [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117740, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.543771] env[63531]: DEBUG nova.network.neutron [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Updating instance_info_cache with network_info: [{"id": "f178c862-e753-40d4-b7cd-3eabe7eda39e", "address": "fa:16:3e:a6:d3:d2", "network": {"id": "5439232a-ec6d-4184-a9f1-f7be5c1526bc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.35", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "54fbef8a6ee0488f820706d2036e9c99", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f267bcdd-0daa-4337-9709-5fc060c267d8", "external-id": "nsx-vlan-transportzone-308", "segmentation_id": 308, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf178c862-e7", "ovs_interfaceid": "f178c862-e753-40d4-b7cd-3eabe7eda39e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 780.570770] env[63531]: DEBUG oslo_vmware.api [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117736, 'name': ReconfigVM_Task, 'duration_secs': 1.432308} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.574700] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Reconfigured VM instance instance-00000032 to attach disk [datastore1] 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f/2998fb3d-6677-4fb3-952a-c74d8c6b7b9f.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 780.575688] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6ba27cf6-f1b3-48dc-8cfc-a942974d6de8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.584323] env[63531]: DEBUG oslo_vmware.api [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Waiting for the task: (returnval){ [ 780.584323] env[63531]: value = "task-1117741" [ 780.584323] env[63531]: _type = "Task" [ 780.584323] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.598199] env[63531]: DEBUG oslo_vmware.api [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117741, 'name': Rename_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.601305] env[63531]: DEBUG nova.compute.manager [req-ed61bdb4-8e51-4ece-96b5-2b36e4ec8051 req-acf77cf6-3df6-4bae-83fc-8dfca60ae0ce service nova] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Received event network-vif-plugged-b4141cf4-dab5-4033-8cc5-06966340323b {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 780.601623] env[63531]: DEBUG oslo_concurrency.lockutils [req-ed61bdb4-8e51-4ece-96b5-2b36e4ec8051 req-acf77cf6-3df6-4bae-83fc-8dfca60ae0ce service nova] Acquiring lock "0197e03d-6c36-4e73-9472-8e3319eb89ce-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 780.601942] env[63531]: DEBUG oslo_concurrency.lockutils [req-ed61bdb4-8e51-4ece-96b5-2b36e4ec8051 req-acf77cf6-3df6-4bae-83fc-8dfca60ae0ce service nova] Lock "0197e03d-6c36-4e73-9472-8e3319eb89ce-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 780.602205] env[63531]: DEBUG oslo_concurrency.lockutils [req-ed61bdb4-8e51-4ece-96b5-2b36e4ec8051 req-acf77cf6-3df6-4bae-83fc-8dfca60ae0ce service nova] Lock "0197e03d-6c36-4e73-9472-8e3319eb89ce-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 780.602467] env[63531]: DEBUG nova.compute.manager [req-ed61bdb4-8e51-4ece-96b5-2b36e4ec8051 req-acf77cf6-3df6-4bae-83fc-8dfca60ae0ce service nova] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] No waiting events found dispatching network-vif-plugged-b4141cf4-dab5-4033-8cc5-06966340323b {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 780.602733] env[63531]: WARNING nova.compute.manager [req-ed61bdb4-8e51-4ece-96b5-2b36e4ec8051 req-acf77cf6-3df6-4bae-83fc-8dfca60ae0ce service nova] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Received unexpected event network-vif-plugged-b4141cf4-dab5-4033-8cc5-06966340323b for instance with vm_state building and task_state spawning. [ 780.603052] env[63531]: DEBUG nova.compute.manager [req-ed61bdb4-8e51-4ece-96b5-2b36e4ec8051 req-acf77cf6-3df6-4bae-83fc-8dfca60ae0ce service nova] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Received event network-changed-b4141cf4-dab5-4033-8cc5-06966340323b {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 780.603354] env[63531]: DEBUG nova.compute.manager [req-ed61bdb4-8e51-4ece-96b5-2b36e4ec8051 req-acf77cf6-3df6-4bae-83fc-8dfca60ae0ce service nova] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Refreshing instance network info cache due to event network-changed-b4141cf4-dab5-4033-8cc5-06966340323b. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 780.603804] env[63531]: DEBUG oslo_concurrency.lockutils [req-ed61bdb4-8e51-4ece-96b5-2b36e4ec8051 req-acf77cf6-3df6-4bae-83fc-8dfca60ae0ce service nova] Acquiring lock "refresh_cache-0197e03d-6c36-4e73-9472-8e3319eb89ce" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 780.659712] env[63531]: DEBUG nova.network.neutron [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Updating instance_info_cache with network_info: [{"id": "b4141cf4-dab5-4033-8cc5-06966340323b", "address": "fa:16:3e:72:a7:59", "network": {"id": "7af8bbee-a43a-40b5-99c4-aee7e19b4093", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1560507257-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "68eaa455a44b4d96936fe8273629279a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ed82fc44-f216-40a7-a68b-62d76ffdb1f8", "external-id": "nsx-vlan-transportzone-236", "segmentation_id": 236, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4141cf4-da", "ovs_interfaceid": "b4141cf4-dab5-4033-8cc5-06966340323b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 780.748696] env[63531]: DEBUG nova.compute.manager [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 780.781283] env[63531]: DEBUG nova.virt.hardware [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 780.782033] env[63531]: DEBUG nova.virt.hardware [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 780.782033] env[63531]: DEBUG nova.virt.hardware [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 780.782033] env[63531]: DEBUG nova.virt.hardware [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 780.782033] env[63531]: DEBUG nova.virt.hardware [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 780.782475] env[63531]: DEBUG nova.virt.hardware [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 780.782475] env[63531]: DEBUG nova.virt.hardware [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 780.782586] env[63531]: DEBUG nova.virt.hardware [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 780.783036] env[63531]: DEBUG nova.virt.hardware [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 780.783036] env[63531]: DEBUG nova.virt.hardware [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 780.783168] env[63531]: DEBUG nova.virt.hardware [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 780.784083] env[63531]: DEBUG nova.scheduler.client.report [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 780.788610] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-029a42bd-8e27-4f01-a7a8-d34b614a51ba {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.798900] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ba9c4ca-1788-40a4-bd30-99d3c822a4a3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.837593] env[63531]: DEBUG nova.compute.manager [req-667f9321-c850-4939-a7a3-a2bb4118f4be req-203ca77e-fe96-4e4b-a140-1c1de01518cf service nova] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Received event network-changed-76060df9-1ff6-4c64-86ef-a8e0d4081cfd {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 780.837804] env[63531]: DEBUG nova.compute.manager [req-667f9321-c850-4939-a7a3-a2bb4118f4be req-203ca77e-fe96-4e4b-a140-1c1de01518cf service nova] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Refreshing instance network info cache due to event network-changed-76060df9-1ff6-4c64-86ef-a8e0d4081cfd. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 780.838340] env[63531]: DEBUG oslo_concurrency.lockutils [req-667f9321-c850-4939-a7a3-a2bb4118f4be req-203ca77e-fe96-4e4b-a140-1c1de01518cf service nova] Acquiring lock "refresh_cache-8dec0cac-defc-42f2-ab56-6b3ae60ad858" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 780.838340] env[63531]: DEBUG oslo_concurrency.lockutils [req-667f9321-c850-4939-a7a3-a2bb4118f4be req-203ca77e-fe96-4e4b-a140-1c1de01518cf service nova] Acquired lock "refresh_cache-8dec0cac-defc-42f2-ab56-6b3ae60ad858" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.838340] env[63531]: DEBUG nova.network.neutron [req-667f9321-c850-4939-a7a3-a2bb4118f4be req-203ca77e-fe96-4e4b-a140-1c1de01518cf service nova] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Refreshing network info cache for port 76060df9-1ff6-4c64-86ef-a8e0d4081cfd {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 780.853120] env[63531]: DEBUG oslo_vmware.api [None req-637af179-656f-4b75-bdcc-db626dfdd5bd tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Task: {'id': task-1117739, 'name': ReconfigVM_Task, 'duration_secs': 0.245569} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.856322] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-637af179-656f-4b75-bdcc-db626dfdd5bd tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Reconfigured VM instance instance-00000029 to detach disk 2000 {{(pid=63531) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 780.860763] env[63531]: DEBUG oslo_vmware.api [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117740, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.484382} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.862317] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-82140846-6df4-47e1-8bb3-95897f53df14 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.871016] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 64fb5831-0789-482b-bb98-67b29868c4c9/64fb5831-0789-482b-bb98-67b29868c4c9.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 780.871170] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 780.871964] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-126da125-c074-4db6-991f-9e5eddfd641d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.879495] env[63531]: DEBUG oslo_vmware.api [None req-637af179-656f-4b75-bdcc-db626dfdd5bd tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Waiting for the task: (returnval){ [ 780.879495] env[63531]: value = "task-1117742" [ 780.879495] env[63531]: _type = "Task" [ 780.879495] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.881149] env[63531]: DEBUG oslo_vmware.api [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Waiting for the task: (returnval){ [ 780.881149] env[63531]: value = "task-1117743" [ 780.881149] env[63531]: _type = "Task" [ 780.881149] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.893090] env[63531]: DEBUG oslo_vmware.api [None req-637af179-656f-4b75-bdcc-db626dfdd5bd tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Task: {'id': task-1117742, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.896194] env[63531]: DEBUG oslo_vmware.api [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117743, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.046739] env[63531]: DEBUG oslo_concurrency.lockutils [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Releasing lock "refresh_cache-5e9042b3-4dc2-4fa3-a664-c4b49a22e400" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 781.094679] env[63531]: DEBUG oslo_vmware.api [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117741, 'name': Rename_Task, 'duration_secs': 0.184029} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.094999] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 781.095285] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-debeb893-9403-45e6-9538-31b6c2a191a9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.101297] env[63531]: DEBUG oslo_vmware.api [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Waiting for the task: (returnval){ [ 781.101297] env[63531]: value = "task-1117744" [ 781.101297] env[63531]: _type = "Task" [ 781.101297] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.110958] env[63531]: DEBUG oslo_vmware.api [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117744, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.165490] env[63531]: DEBUG oslo_concurrency.lockutils [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Releasing lock "refresh_cache-0197e03d-6c36-4e73-9472-8e3319eb89ce" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 781.165887] env[63531]: DEBUG nova.compute.manager [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Instance network_info: |[{"id": "b4141cf4-dab5-4033-8cc5-06966340323b", "address": "fa:16:3e:72:a7:59", "network": {"id": "7af8bbee-a43a-40b5-99c4-aee7e19b4093", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1560507257-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "68eaa455a44b4d96936fe8273629279a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ed82fc44-f216-40a7-a68b-62d76ffdb1f8", "external-id": "nsx-vlan-transportzone-236", "segmentation_id": 236, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4141cf4-da", "ovs_interfaceid": "b4141cf4-dab5-4033-8cc5-06966340323b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 781.166265] env[63531]: DEBUG oslo_concurrency.lockutils [req-ed61bdb4-8e51-4ece-96b5-2b36e4ec8051 req-acf77cf6-3df6-4bae-83fc-8dfca60ae0ce service nova] Acquired lock "refresh_cache-0197e03d-6c36-4e73-9472-8e3319eb89ce" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.166460] env[63531]: DEBUG nova.network.neutron [req-ed61bdb4-8e51-4ece-96b5-2b36e4ec8051 req-acf77cf6-3df6-4bae-83fc-8dfca60ae0ce service nova] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Refreshing network info cache for port b4141cf4-dab5-4033-8cc5-06966340323b {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 781.167702] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:72:a7:59', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ed82fc44-f216-40a7-a68b-62d76ffdb1f8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b4141cf4-dab5-4033-8cc5-06966340323b', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 781.176177] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Creating folder: Project (68eaa455a44b4d96936fe8273629279a). Parent ref: group-v244585. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 781.179621] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e0ab239b-4a07-4d0a-985b-9a1912dbe2a6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.193025] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Created folder: Project (68eaa455a44b4d96936fe8273629279a) in parent group-v244585. [ 781.193025] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Creating folder: Instances. Parent ref: group-v244656. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 781.193025] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1525efd7-566b-4369-859a-4783aa02ef22 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.203823] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Created folder: Instances in parent group-v244656. [ 781.203823] env[63531]: DEBUG oslo.service.loopingcall [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 781.203823] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 781.203985] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-453f2598-68f0-4bb9-a083-d7ef64df1675 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.228037] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 781.228037] env[63531]: value = "task-1117747" [ 781.228037] env[63531]: _type = "Task" [ 781.228037] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.239836] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117747, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.289360] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.579s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.289914] env[63531]: DEBUG nova.compute.manager [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 781.294011] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a56197e5-b9a2-4ba5-a23f-e676b6cc2b13 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.307s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.294011] env[63531]: DEBUG nova.objects.instance [None req-a56197e5-b9a2-4ba5-a23f-e676b6cc2b13 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Lazy-loading 'resources' on Instance uuid 0af99240-fa7a-4eed-a729-a5ae98c41cf9 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 781.397799] env[63531]: DEBUG oslo_vmware.api [None req-637af179-656f-4b75-bdcc-db626dfdd5bd tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Task: {'id': task-1117742, 'name': ReconfigVM_Task, 'duration_secs': 0.104659} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.397799] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-637af179-656f-4b75-bdcc-db626dfdd5bd tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244601', 'volume_id': '2a24c282-f681-40bf-a19c-5c711d342687', 'name': 'volume-2a24c282-f681-40bf-a19c-5c711d342687', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '410ddf7a-e45b-4769-a3db-5363ce2096a7', 'attached_at': '', 'detached_at': '', 'volume_id': '2a24c282-f681-40bf-a19c-5c711d342687', 'serial': '2a24c282-f681-40bf-a19c-5c711d342687'} {{(pid=63531) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 781.397799] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-637af179-656f-4b75-bdcc-db626dfdd5bd tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 781.398078] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2671cd6b-ae81-44b7-ac9b-3d960c2e0a60 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.404180] env[63531]: DEBUG oslo_vmware.api [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117743, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07491} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.404180] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 781.404637] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2ceab09-409e-412c-b87d-e491df1633c9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.408948] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-637af179-656f-4b75-bdcc-db626dfdd5bd tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 781.409506] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-671a44fb-3b43-4440-9dcc-d595c1425a11 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.428875] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Reconfiguring VM instance instance-00000033 to attach disk [datastore1] 64fb5831-0789-482b-bb98-67b29868c4c9/64fb5831-0789-482b-bb98-67b29868c4c9.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 781.428875] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-53b4fa17-f428-45fa-ab0b-65ff22fc8bdf {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.450669] env[63531]: DEBUG oslo_vmware.api [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Waiting for the task: (returnval){ [ 781.450669] env[63531]: value = "task-1117749" [ 781.450669] env[63531]: _type = "Task" [ 781.450669] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.459538] env[63531]: DEBUG oslo_vmware.api [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117749, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.468542] env[63531]: DEBUG nova.network.neutron [req-ed61bdb4-8e51-4ece-96b5-2b36e4ec8051 req-acf77cf6-3df6-4bae-83fc-8dfca60ae0ce service nova] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Updated VIF entry in instance network info cache for port b4141cf4-dab5-4033-8cc5-06966340323b. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 781.469116] env[63531]: DEBUG nova.network.neutron [req-ed61bdb4-8e51-4ece-96b5-2b36e4ec8051 req-acf77cf6-3df6-4bae-83fc-8dfca60ae0ce service nova] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Updating instance_info_cache with network_info: [{"id": "b4141cf4-dab5-4033-8cc5-06966340323b", "address": "fa:16:3e:72:a7:59", "network": {"id": "7af8bbee-a43a-40b5-99c4-aee7e19b4093", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1560507257-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "68eaa455a44b4d96936fe8273629279a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ed82fc44-f216-40a7-a68b-62d76ffdb1f8", "external-id": "nsx-vlan-transportzone-236", "segmentation_id": 236, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4141cf4-da", "ovs_interfaceid": "b4141cf4-dab5-4033-8cc5-06966340323b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.479031] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-637af179-656f-4b75-bdcc-db626dfdd5bd tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 781.479031] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-637af179-656f-4b75-bdcc-db626dfdd5bd tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 781.479031] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-637af179-656f-4b75-bdcc-db626dfdd5bd tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Deleting the datastore file [datastore1] 410ddf7a-e45b-4769-a3db-5363ce2096a7 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 781.479031] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e43e1d4d-9171-49f7-b3b6-3e289d78dea2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.484353] env[63531]: DEBUG oslo_vmware.api [None req-637af179-656f-4b75-bdcc-db626dfdd5bd tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Waiting for the task: (returnval){ [ 781.484353] env[63531]: value = "task-1117750" [ 781.484353] env[63531]: _type = "Task" [ 781.484353] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.494913] env[63531]: DEBUG oslo_vmware.api [None req-637af179-656f-4b75-bdcc-db626dfdd5bd tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Task: {'id': task-1117750, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.576438] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b559e631-780e-40be-9b72-6a2125bfa3c1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.598107] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33dd5bcd-48b0-41f2-94bb-8cb07039857e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.608568] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Updating instance '5e9042b3-4dc2-4fa3-a664-c4b49a22e400' progress to 83 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 781.617889] env[63531]: DEBUG oslo_vmware.api [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117744, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.624966] env[63531]: DEBUG nova.network.neutron [req-667f9321-c850-4939-a7a3-a2bb4118f4be req-203ca77e-fe96-4e4b-a140-1c1de01518cf service nova] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Updated VIF entry in instance network info cache for port 76060df9-1ff6-4c64-86ef-a8e0d4081cfd. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 781.625358] env[63531]: DEBUG nova.network.neutron [req-667f9321-c850-4939-a7a3-a2bb4118f4be req-203ca77e-fe96-4e4b-a140-1c1de01518cf service nova] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Updating instance_info_cache with network_info: [{"id": "76060df9-1ff6-4c64-86ef-a8e0d4081cfd", "address": "fa:16:3e:52:f5:96", "network": {"id": "eac8786d-bc67-4916-8483-b53d9a25efb8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1604989690-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.241", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "abdae960d43b47bc95ab41f265d2c85b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db068f71-08cc-42d4-8ab6-17134c1585e5", "external-id": "nsx-vlan-transportzone-721", "segmentation_id": 721, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap76060df9-1f", "ovs_interfaceid": "76060df9-1ff6-4c64-86ef-a8e0d4081cfd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.737349] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117747, 'name': CreateVM_Task, 'duration_secs': 0.378592} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.737523] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 781.738234] env[63531]: DEBUG oslo_concurrency.lockutils [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 781.738404] env[63531]: DEBUG oslo_concurrency.lockutils [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.738724] env[63531]: DEBUG oslo_concurrency.lockutils [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 781.738971] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2a7dfadf-543e-43c1-8be4-39f7a57a5452 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.743292] env[63531]: DEBUG oslo_vmware.api [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Waiting for the task: (returnval){ [ 781.743292] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]525896d0-bc30-1374-82ec-9775e0648104" [ 781.743292] env[63531]: _type = "Task" [ 781.743292] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.752587] env[63531]: DEBUG oslo_vmware.api [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]525896d0-bc30-1374-82ec-9775e0648104, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.800752] env[63531]: DEBUG nova.compute.utils [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 781.803185] env[63531]: DEBUG nova.compute.manager [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 781.803185] env[63531]: DEBUG nova.network.neutron [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 781.853842] env[63531]: DEBUG nova.policy [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2ec68c1860c5484cb8094370baa80c98', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '57e9bf4ee0f1463fa67b0f8deff30fdc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 781.962535] env[63531]: DEBUG oslo_vmware.api [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117749, 'name': ReconfigVM_Task, 'duration_secs': 0.289735} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.965157] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Reconfigured VM instance instance-00000033 to attach disk [datastore1] 64fb5831-0789-482b-bb98-67b29868c4c9/64fb5831-0789-482b-bb98-67b29868c4c9.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 781.965991] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6cc1737f-3340-4f7c-b3ec-cfe8bc5f9e1b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.973501] env[63531]: DEBUG oslo_concurrency.lockutils [req-ed61bdb4-8e51-4ece-96b5-2b36e4ec8051 req-acf77cf6-3df6-4bae-83fc-8dfca60ae0ce service nova] Releasing lock "refresh_cache-0197e03d-6c36-4e73-9472-8e3319eb89ce" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 781.973925] env[63531]: DEBUG oslo_vmware.api [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Waiting for the task: (returnval){ [ 781.973925] env[63531]: value = "task-1117751" [ 781.973925] env[63531]: _type = "Task" [ 781.973925] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.984126] env[63531]: DEBUG oslo_vmware.api [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117751, 'name': Rename_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.993657] env[63531]: DEBUG oslo_vmware.api [None req-637af179-656f-4b75-bdcc-db626dfdd5bd tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Task: {'id': task-1117750, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.084874} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.997059] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-637af179-656f-4b75-bdcc-db626dfdd5bd tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 781.997304] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-637af179-656f-4b75-bdcc-db626dfdd5bd tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 781.997509] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-637af179-656f-4b75-bdcc-db626dfdd5bd tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 782.035608] env[63531]: DEBUG nova.network.neutron [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Successfully updated port: d4b7599b-ca73-4bab-91d4-33a977545afc {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 782.072069] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-637af179-656f-4b75-bdcc-db626dfdd5bd tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Volume detach. Driver type: vmdk {{(pid=63531) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 782.072407] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-da64d0ee-122b-46a6-b208-56fe8e5863a0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.083272] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad36f2b8-c995-4ff6-9882-c5941182d897 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.116189] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 782.121047] env[63531]: ERROR nova.compute.manager [None req-637af179-656f-4b75-bdcc-db626dfdd5bd tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Failed to detach volume 2a24c282-f681-40bf-a19c-5c711d342687 from /dev/sda: nova.exception.InstanceNotFound: Instance 410ddf7a-e45b-4769-a3db-5363ce2096a7 could not be found. [ 782.121047] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Traceback (most recent call last): [ 782.121047] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] File "/opt/stack/nova/nova/compute/manager.py", line 4142, in _do_rebuild_instance [ 782.121047] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] self.driver.rebuild(**kwargs) [ 782.121047] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] File "/opt/stack/nova/nova/virt/driver.py", line 493, in rebuild [ 782.121047] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] raise NotImplementedError() [ 782.121047] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] NotImplementedError [ 782.121047] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] [ 782.121047] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] During handling of the above exception, another exception occurred: [ 782.121047] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] [ 782.121047] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Traceback (most recent call last): [ 782.121047] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] File "/opt/stack/nova/nova/compute/manager.py", line 3565, in _detach_root_volume [ 782.121047] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] self.driver.detach_volume(context, old_connection_info, [ 782.121047] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 553, in detach_volume [ 782.121047] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] return self._volumeops.detach_volume(connection_info, instance) [ 782.121047] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 782.121047] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] self._detach_volume_vmdk(connection_info, instance) [ 782.121047] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 782.121047] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 782.121047] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1135, in get_vm_ref [ 782.121047] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] stable_ref.fetch_moref(session) [ 782.121047] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1126, in fetch_moref [ 782.121047] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] raise exception.InstanceNotFound(instance_id=self._uuid) [ 782.121047] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] nova.exception.InstanceNotFound: Instance 410ddf7a-e45b-4769-a3db-5363ce2096a7 could not be found. [ 782.121047] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] [ 782.127114] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d9998fb8-059c-4e84-aef2-9d689ad0d790 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.128261] env[63531]: DEBUG oslo_vmware.api [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117744, 'name': PowerOnVM_Task, 'duration_secs': 0.56266} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.129497] env[63531]: DEBUG oslo_concurrency.lockutils [req-667f9321-c850-4939-a7a3-a2bb4118f4be req-203ca77e-fe96-4e4b-a140-1c1de01518cf service nova] Releasing lock "refresh_cache-8dec0cac-defc-42f2-ab56-6b3ae60ad858" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.129497] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 782.129812] env[63531]: INFO nova.compute.manager [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Took 9.23 seconds to spawn the instance on the hypervisor. [ 782.130061] env[63531]: DEBUG nova.compute.manager [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 782.131276] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c18d0d93-1a74-47b7-b04b-1adffc31c568 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.135538] env[63531]: DEBUG oslo_vmware.api [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Waiting for the task: (returnval){ [ 782.135538] env[63531]: value = "task-1117752" [ 782.135538] env[63531]: _type = "Task" [ 782.135538] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.152996] env[63531]: DEBUG oslo_vmware.api [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Task: {'id': task-1117752, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.215317] env[63531]: DEBUG nova.network.neutron [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Successfully created port: e4133e0f-55c1-4b3f-847d-899dd1167eec {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 782.255823] env[63531]: DEBUG oslo_vmware.api [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]525896d0-bc30-1374-82ec-9775e0648104, 'name': SearchDatastore_Task, 'duration_secs': 0.008508} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.261173] env[63531]: DEBUG oslo_concurrency.lockutils [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.261447] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 782.261682] env[63531]: DEBUG oslo_concurrency.lockutils [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.261841] env[63531]: DEBUG oslo_concurrency.lockutils [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.262047] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 782.263010] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b36bee2e-bd7e-469c-a6bb-2ee2f6a07add {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.272953] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 782.273223] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 782.273937] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-198f25ca-708c-4ce6-8289-01b4c9de6973 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.280137] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ca23fb7-ee00-43a4-bd54-7b29503f7272 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.283264] env[63531]: DEBUG oslo_vmware.api [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Waiting for the task: (returnval){ [ 782.283264] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]524a1a31-b32b-1148-3276-00566ce346af" [ 782.283264] env[63531]: _type = "Task" [ 782.283264] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.289942] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30a4eb18-8274-44e4-9820-207630741072 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.298799] env[63531]: DEBUG oslo_vmware.api [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]524a1a31-b32b-1148-3276-00566ce346af, 'name': SearchDatastore_Task, 'duration_secs': 0.008218} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.299893] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-84269ff7-c6f0-4faf-877a-8ba8fbdfe38b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.327868] env[63531]: DEBUG nova.compute.manager [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 782.333823] env[63531]: DEBUG nova.compute.utils [None req-637af179-656f-4b75-bdcc-db626dfdd5bd tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Build of instance 410ddf7a-e45b-4769-a3db-5363ce2096a7 aborted: Failed to rebuild volume backed instance. {{(pid=63531) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 782.334475] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3021fc0-d077-400a-b868-be3bda576024 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.338668] env[63531]: DEBUG oslo_vmware.api [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Waiting for the task: (returnval){ [ 782.338668] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5250afd5-2bf0-789d-3d49-0b76fad63101" [ 782.338668] env[63531]: _type = "Task" [ 782.338668] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.345796] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc6e818b-e89b-429a-80b5-e94374c0e1d6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.351128] env[63531]: ERROR nova.compute.manager [None req-637af179-656f-4b75-bdcc-db626dfdd5bd tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Setting instance vm_state to ERROR: nova.exception.BuildAbortException: Build of instance 410ddf7a-e45b-4769-a3db-5363ce2096a7 aborted: Failed to rebuild volume backed instance. [ 782.351128] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Traceback (most recent call last): [ 782.351128] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] File "/opt/stack/nova/nova/compute/manager.py", line 4142, in _do_rebuild_instance [ 782.351128] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] self.driver.rebuild(**kwargs) [ 782.351128] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] File "/opt/stack/nova/nova/virt/driver.py", line 493, in rebuild [ 782.351128] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] raise NotImplementedError() [ 782.351128] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] NotImplementedError [ 782.351128] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] [ 782.351128] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] During handling of the above exception, another exception occurred: [ 782.351128] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] [ 782.351128] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Traceback (most recent call last): [ 782.351128] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] File "/opt/stack/nova/nova/compute/manager.py", line 3600, in _rebuild_volume_backed_instance [ 782.351128] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] self._detach_root_volume(context, instance, root_bdm) [ 782.351128] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] File "/opt/stack/nova/nova/compute/manager.py", line 3579, in _detach_root_volume [ 782.351128] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] with excutils.save_and_reraise_exception(): [ 782.351128] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 782.351128] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] self.force_reraise() [ 782.351128] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 782.351128] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] raise self.value [ 782.351128] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] File "/opt/stack/nova/nova/compute/manager.py", line 3565, in _detach_root_volume [ 782.351128] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] self.driver.detach_volume(context, old_connection_info, [ 782.351128] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 553, in detach_volume [ 782.351128] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] return self._volumeops.detach_volume(connection_info, instance) [ 782.351128] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 782.351128] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] self._detach_volume_vmdk(connection_info, instance) [ 782.351128] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 782.351128] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 782.351128] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1135, in get_vm_ref [ 782.351128] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] stable_ref.fetch_moref(session) [ 782.351128] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1126, in fetch_moref [ 782.351128] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] raise exception.InstanceNotFound(instance_id=self._uuid) [ 782.351128] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] nova.exception.InstanceNotFound: Instance 410ddf7a-e45b-4769-a3db-5363ce2096a7 could not be found. [ 782.351128] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] [ 782.351128] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] During handling of the above exception, another exception occurred: [ 782.351128] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] [ 782.351128] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Traceback (most recent call last): [ 782.351128] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] File "/opt/stack/nova/nova/compute/manager.py", line 10865, in _error_out_instance_on_exception [ 782.351128] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] yield [ 782.351128] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] File "/opt/stack/nova/nova/compute/manager.py", line 3868, in rebuild_instance [ 782.351128] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] self._do_rebuild_instance_with_claim( [ 782.352366] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] File "/opt/stack/nova/nova/compute/manager.py", line 3954, in _do_rebuild_instance_with_claim [ 782.352366] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] self._do_rebuild_instance( [ 782.352366] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] File "/opt/stack/nova/nova/compute/manager.py", line 4146, in _do_rebuild_instance [ 782.352366] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] self._rebuild_default_impl(**kwargs) [ 782.352366] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] File "/opt/stack/nova/nova/compute/manager.py", line 3723, in _rebuild_default_impl [ 782.352366] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] self._rebuild_volume_backed_instance( [ 782.352366] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] File "/opt/stack/nova/nova/compute/manager.py", line 3615, in _rebuild_volume_backed_instance [ 782.352366] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] raise exception.BuildAbortException( [ 782.352366] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] nova.exception.BuildAbortException: Build of instance 410ddf7a-e45b-4769-a3db-5363ce2096a7 aborted: Failed to rebuild volume backed instance. [ 782.352366] env[63531]: ERROR nova.compute.manager [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] [ 782.356984] env[63531]: DEBUG oslo_vmware.api [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5250afd5-2bf0-789d-3d49-0b76fad63101, 'name': SearchDatastore_Task, 'duration_secs': 0.0089} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.357790] env[63531]: DEBUG oslo_concurrency.lockutils [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.358181] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 0197e03d-6c36-4e73-9472-8e3319eb89ce/0197e03d-6c36-4e73-9472-8e3319eb89ce.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 782.358409] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-702247f9-ca29-49bd-8061-0b52efcaac70 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.369237] env[63531]: DEBUG nova.compute.provider_tree [None req-a56197e5-b9a2-4ba5-a23f-e676b6cc2b13 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 782.375513] env[63531]: DEBUG oslo_vmware.api [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Waiting for the task: (returnval){ [ 782.375513] env[63531]: value = "task-1117753" [ 782.375513] env[63531]: _type = "Task" [ 782.375513] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.384559] env[63531]: DEBUG oslo_vmware.api [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Task: {'id': task-1117753, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.483850] env[63531]: DEBUG oslo_vmware.api [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117751, 'name': Rename_Task, 'duration_secs': 0.14794} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.484145] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 782.484394] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0676a9ba-510f-4cb5-8eff-09808375ec8e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.490840] env[63531]: DEBUG oslo_vmware.api [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Waiting for the task: (returnval){ [ 782.490840] env[63531]: value = "task-1117754" [ 782.490840] env[63531]: _type = "Task" [ 782.490840] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.499628] env[63531]: DEBUG oslo_vmware.api [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117754, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.539277] env[63531]: DEBUG oslo_concurrency.lockutils [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Acquiring lock "refresh_cache-1e5e28ef-38ce-47be-ac36-8bdf17ab0a47" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.539453] env[63531]: DEBUG oslo_concurrency.lockutils [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Acquired lock "refresh_cache-1e5e28ef-38ce-47be-ac36-8bdf17ab0a47" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.539623] env[63531]: DEBUG nova.network.neutron [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 782.623729] env[63531]: DEBUG nova.compute.manager [req-421ffca7-f1d8-433b-9ca7-5d0c52b68cd0 req-1e64e7a3-e4ab-49e1-abb5-9d24b20aa328 service nova] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Received event network-vif-plugged-d4b7599b-ca73-4bab-91d4-33a977545afc {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 782.624213] env[63531]: DEBUG oslo_concurrency.lockutils [req-421ffca7-f1d8-433b-9ca7-5d0c52b68cd0 req-1e64e7a3-e4ab-49e1-abb5-9d24b20aa328 service nova] Acquiring lock "1e5e28ef-38ce-47be-ac36-8bdf17ab0a47-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 782.624617] env[63531]: DEBUG oslo_concurrency.lockutils [req-421ffca7-f1d8-433b-9ca7-5d0c52b68cd0 req-1e64e7a3-e4ab-49e1-abb5-9d24b20aa328 service nova] Lock "1e5e28ef-38ce-47be-ac36-8bdf17ab0a47-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 782.624999] env[63531]: DEBUG oslo_concurrency.lockutils [req-421ffca7-f1d8-433b-9ca7-5d0c52b68cd0 req-1e64e7a3-e4ab-49e1-abb5-9d24b20aa328 service nova] Lock "1e5e28ef-38ce-47be-ac36-8bdf17ab0a47-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 782.625421] env[63531]: DEBUG nova.compute.manager [req-421ffca7-f1d8-433b-9ca7-5d0c52b68cd0 req-1e64e7a3-e4ab-49e1-abb5-9d24b20aa328 service nova] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] No waiting events found dispatching network-vif-plugged-d4b7599b-ca73-4bab-91d4-33a977545afc {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 782.625755] env[63531]: WARNING nova.compute.manager [req-421ffca7-f1d8-433b-9ca7-5d0c52b68cd0 req-1e64e7a3-e4ab-49e1-abb5-9d24b20aa328 service nova] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Received unexpected event network-vif-plugged-d4b7599b-ca73-4bab-91d4-33a977545afc for instance with vm_state building and task_state spawning. [ 782.626094] env[63531]: DEBUG nova.compute.manager [req-421ffca7-f1d8-433b-9ca7-5d0c52b68cd0 req-1e64e7a3-e4ab-49e1-abb5-9d24b20aa328 service nova] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Received event network-changed-d4b7599b-ca73-4bab-91d4-33a977545afc {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 782.626389] env[63531]: DEBUG nova.compute.manager [req-421ffca7-f1d8-433b-9ca7-5d0c52b68cd0 req-1e64e7a3-e4ab-49e1-abb5-9d24b20aa328 service nova] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Refreshing instance network info cache due to event network-changed-d4b7599b-ca73-4bab-91d4-33a977545afc. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 782.626741] env[63531]: DEBUG oslo_concurrency.lockutils [req-421ffca7-f1d8-433b-9ca7-5d0c52b68cd0 req-1e64e7a3-e4ab-49e1-abb5-9d24b20aa328 service nova] Acquiring lock "refresh_cache-1e5e28ef-38ce-47be-ac36-8bdf17ab0a47" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.659987] env[63531]: DEBUG oslo_vmware.api [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Task: {'id': task-1117752, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.663663] env[63531]: INFO nova.compute.manager [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Took 40.78 seconds to build instance. [ 782.875478] env[63531]: DEBUG nova.scheduler.client.report [None req-a56197e5-b9a2-4ba5-a23f-e676b6cc2b13 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 782.890036] env[63531]: DEBUG oslo_vmware.api [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Task: {'id': task-1117753, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.487138} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.890616] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 0197e03d-6c36-4e73-9472-8e3319eb89ce/0197e03d-6c36-4e73-9472-8e3319eb89ce.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 782.891034] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 782.891606] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a909a79f-d84b-4943-941d-039fffb984ea {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.899099] env[63531]: DEBUG oslo_vmware.api [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Waiting for the task: (returnval){ [ 782.899099] env[63531]: value = "task-1117755" [ 782.899099] env[63531]: _type = "Task" [ 782.899099] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.908210] env[63531]: DEBUG oslo_vmware.api [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Task: {'id': task-1117755, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.000929] env[63531]: DEBUG oslo_vmware.api [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117754, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.091049] env[63531]: DEBUG nova.network.neutron [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 783.148717] env[63531]: DEBUG oslo_vmware.api [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Task: {'id': task-1117752, 'name': PowerOnVM_Task, 'duration_secs': 0.763867} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.151442] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 783.151720] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-041c9f5b-3b0e-43fe-b06c-9a9afcbf0147 tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Updating instance '5e9042b3-4dc2-4fa3-a664-c4b49a22e400' progress to 100 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 783.167620] env[63531]: DEBUG oslo_concurrency.lockutils [None req-83c76a8b-b236-4f0b-9693-818ac9ec45be tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Lock "2998fb3d-6677-4fb3-952a-c74d8c6b7b9f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 126.095s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.341179] env[63531]: DEBUG nova.compute.manager [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 783.344751] env[63531]: DEBUG nova.network.neutron [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Updating instance_info_cache with network_info: [{"id": "d4b7599b-ca73-4bab-91d4-33a977545afc", "address": "fa:16:3e:5e:c6:28", "network": {"id": "887fff70-ea19-458d-a945-e4f3ed6dc185", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2099566288-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "04a4c814245346a9b73253a25a822fef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49a4d142-3f97-47fe-b074-58923c46815e", "external-id": "nsx-vlan-transportzone-565", "segmentation_id": 565, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd4b7599b-ca", "ovs_interfaceid": "d4b7599b-ca73-4bab-91d4-33a977545afc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.381956] env[63531]: DEBUG nova.virt.hardware [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 783.382226] env[63531]: DEBUG nova.virt.hardware [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 783.382387] env[63531]: DEBUG nova.virt.hardware [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 783.382571] env[63531]: DEBUG nova.virt.hardware [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 783.382717] env[63531]: DEBUG nova.virt.hardware [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 783.382863] env[63531]: DEBUG nova.virt.hardware [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 783.383940] env[63531]: DEBUG nova.virt.hardware [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 783.383940] env[63531]: DEBUG nova.virt.hardware [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 783.383940] env[63531]: DEBUG nova.virt.hardware [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 783.383940] env[63531]: DEBUG nova.virt.hardware [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 783.383940] env[63531]: DEBUG nova.virt.hardware [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 783.385450] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a56197e5-b9a2-4ba5-a23f-e676b6cc2b13 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.092s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.388111] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-742fc175-d1dc-4f9b-b4cc-83470eb7373e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.391164] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.536s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 783.392729] env[63531]: INFO nova.compute.claims [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 783.405178] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66d4f28f-822a-43db-9bc8-752e8317f44b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.415815] env[63531]: DEBUG oslo_vmware.api [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Task: {'id': task-1117755, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069278} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.425572] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 783.427202] env[63531]: INFO nova.scheduler.client.report [None req-a56197e5-b9a2-4ba5-a23f-e676b6cc2b13 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Deleted allocations for instance 0af99240-fa7a-4eed-a729-a5ae98c41cf9 [ 783.428798] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-227c1a2b-56fd-47ee-a942-86d3b9f429f3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.455750] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Reconfiguring VM instance instance-00000034 to attach disk [datastore1] 0197e03d-6c36-4e73-9472-8e3319eb89ce/0197e03d-6c36-4e73-9472-8e3319eb89ce.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 783.456255] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2dbeccd7-4865-4926-8e3f-c1be162859f0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.476202] env[63531]: DEBUG oslo_vmware.api [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Waiting for the task: (returnval){ [ 783.476202] env[63531]: value = "task-1117756" [ 783.476202] env[63531]: _type = "Task" [ 783.476202] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.485180] env[63531]: DEBUG oslo_vmware.api [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Task: {'id': task-1117756, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.502035] env[63531]: DEBUG oslo_vmware.api [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117754, 'name': PowerOnVM_Task, 'duration_secs': 0.763707} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.502181] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 783.502328] env[63531]: INFO nova.compute.manager [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Took 7.87 seconds to spawn the instance on the hypervisor. [ 783.502498] env[63531]: DEBUG nova.compute.manager [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 783.503257] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb7da5b5-cc1a-489f-a29f-cc2f851320c6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.644850] env[63531]: DEBUG nova.compute.manager [req-4a38cc24-19af-4a6d-87fe-db2a59ceeeee req-12e2bcdb-3815-43c3-af79-284128cb8558 service nova] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Received event network-vif-plugged-e4133e0f-55c1-4b3f-847d-899dd1167eec {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 783.645153] env[63531]: DEBUG oslo_concurrency.lockutils [req-4a38cc24-19af-4a6d-87fe-db2a59ceeeee req-12e2bcdb-3815-43c3-af79-284128cb8558 service nova] Acquiring lock "9a3fad46-a15a-451c-bdab-a3c8cc8add07-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 783.645417] env[63531]: DEBUG oslo_concurrency.lockutils [req-4a38cc24-19af-4a6d-87fe-db2a59ceeeee req-12e2bcdb-3815-43c3-af79-284128cb8558 service nova] Lock "9a3fad46-a15a-451c-bdab-a3c8cc8add07-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 783.645540] env[63531]: DEBUG oslo_concurrency.lockutils [req-4a38cc24-19af-4a6d-87fe-db2a59ceeeee req-12e2bcdb-3815-43c3-af79-284128cb8558 service nova] Lock "9a3fad46-a15a-451c-bdab-a3c8cc8add07-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.645707] env[63531]: DEBUG nova.compute.manager [req-4a38cc24-19af-4a6d-87fe-db2a59ceeeee req-12e2bcdb-3815-43c3-af79-284128cb8558 service nova] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] No waiting events found dispatching network-vif-plugged-e4133e0f-55c1-4b3f-847d-899dd1167eec {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 783.645869] env[63531]: WARNING nova.compute.manager [req-4a38cc24-19af-4a6d-87fe-db2a59ceeeee req-12e2bcdb-3815-43c3-af79-284128cb8558 service nova] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Received unexpected event network-vif-plugged-e4133e0f-55c1-4b3f-847d-899dd1167eec for instance with vm_state building and task_state spawning. [ 783.670781] env[63531]: DEBUG nova.compute.manager [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 783.742926] env[63531]: DEBUG nova.network.neutron [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Successfully updated port: e4133e0f-55c1-4b3f-847d-899dd1167eec {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 783.849476] env[63531]: DEBUG oslo_concurrency.lockutils [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Releasing lock "refresh_cache-1e5e28ef-38ce-47be-ac36-8bdf17ab0a47" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 783.849827] env[63531]: DEBUG nova.compute.manager [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Instance network_info: |[{"id": "d4b7599b-ca73-4bab-91d4-33a977545afc", "address": "fa:16:3e:5e:c6:28", "network": {"id": "887fff70-ea19-458d-a945-e4f3ed6dc185", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2099566288-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "04a4c814245346a9b73253a25a822fef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49a4d142-3f97-47fe-b074-58923c46815e", "external-id": "nsx-vlan-transportzone-565", "segmentation_id": 565, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd4b7599b-ca", "ovs_interfaceid": "d4b7599b-ca73-4bab-91d4-33a977545afc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 783.850140] env[63531]: DEBUG oslo_concurrency.lockutils [req-421ffca7-f1d8-433b-9ca7-5d0c52b68cd0 req-1e64e7a3-e4ab-49e1-abb5-9d24b20aa328 service nova] Acquired lock "refresh_cache-1e5e28ef-38ce-47be-ac36-8bdf17ab0a47" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.850318] env[63531]: DEBUG nova.network.neutron [req-421ffca7-f1d8-433b-9ca7-5d0c52b68cd0 req-1e64e7a3-e4ab-49e1-abb5-9d24b20aa328 service nova] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Refreshing network info cache for port d4b7599b-ca73-4bab-91d4-33a977545afc {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 783.851535] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5e:c6:28', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '49a4d142-3f97-47fe-b074-58923c46815e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd4b7599b-ca73-4bab-91d4-33a977545afc', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 783.859354] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Creating folder: Project (04a4c814245346a9b73253a25a822fef). Parent ref: group-v244585. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 783.862886] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-94bb9a98-0626-4fc8-9b5b-edf866f25069 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.873723] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Created folder: Project (04a4c814245346a9b73253a25a822fef) in parent group-v244585. [ 783.873723] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Creating folder: Instances. Parent ref: group-v244659. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 783.874115] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6f22e645-988d-40f1-96ec-610af2fcfa15 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.883846] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Created folder: Instances in parent group-v244659. [ 783.884216] env[63531]: DEBUG oslo.service.loopingcall [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 783.884321] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 783.884525] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d908b7f6-7f6a-45aa-8fe0-f6a6d385cd10 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.907783] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 783.907783] env[63531]: value = "task-1117759" [ 783.907783] env[63531]: _type = "Task" [ 783.907783] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.916838] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117759, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.938522] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a56197e5-b9a2-4ba5-a23f-e676b6cc2b13 tempest-ServerDiagnosticsNegativeTest-263844179 tempest-ServerDiagnosticsNegativeTest-263844179-project-member] Lock "0af99240-fa7a-4eed-a729-a5ae98c41cf9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.937s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.987185] env[63531]: DEBUG oslo_vmware.api [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Task: {'id': task-1117756, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.026177] env[63531]: INFO nova.compute.manager [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Took 41.99 seconds to build instance. [ 784.197048] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.245323] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquiring lock "refresh_cache-9a3fad46-a15a-451c-bdab-a3c8cc8add07" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 784.245561] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquired lock "refresh_cache-9a3fad46-a15a-451c-bdab-a3c8cc8add07" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 784.245679] env[63531]: DEBUG nova.network.neutron [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 784.256917] env[63531]: DEBUG nova.network.neutron [req-421ffca7-f1d8-433b-9ca7-5d0c52b68cd0 req-1e64e7a3-e4ab-49e1-abb5-9d24b20aa328 service nova] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Updated VIF entry in instance network info cache for port d4b7599b-ca73-4bab-91d4-33a977545afc. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 784.257399] env[63531]: DEBUG nova.network.neutron [req-421ffca7-f1d8-433b-9ca7-5d0c52b68cd0 req-1e64e7a3-e4ab-49e1-abb5-9d24b20aa328 service nova] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Updating instance_info_cache with network_info: [{"id": "d4b7599b-ca73-4bab-91d4-33a977545afc", "address": "fa:16:3e:5e:c6:28", "network": {"id": "887fff70-ea19-458d-a945-e4f3ed6dc185", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2099566288-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "04a4c814245346a9b73253a25a822fef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49a4d142-3f97-47fe-b074-58923c46815e", "external-id": "nsx-vlan-transportzone-565", "segmentation_id": 565, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd4b7599b-ca", "ovs_interfaceid": "d4b7599b-ca73-4bab-91d4-33a977545afc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 784.374575] env[63531]: DEBUG oslo_concurrency.lockutils [None req-637af179-656f-4b75-bdcc-db626dfdd5bd tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.419249] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117759, 'name': CreateVM_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.487719] env[63531]: DEBUG oslo_vmware.api [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Task: {'id': task-1117756, 'name': ReconfigVM_Task, 'duration_secs': 0.770116} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.488112] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Reconfigured VM instance instance-00000034 to attach disk [datastore1] 0197e03d-6c36-4e73-9472-8e3319eb89ce/0197e03d-6c36-4e73-9472-8e3319eb89ce.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 784.488866] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-88e2eebb-e936-4d37-836a-fd2f676bcda6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.497428] env[63531]: DEBUG oslo_vmware.api [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Waiting for the task: (returnval){ [ 784.497428] env[63531]: value = "task-1117760" [ 784.497428] env[63531]: _type = "Task" [ 784.497428] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.507318] env[63531]: DEBUG oslo_vmware.api [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Task: {'id': task-1117760, 'name': Rename_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.531424] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f813de60-369f-4cc7-9345-25df2dc60930 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Lock "64fb5831-0789-482b-bb98-67b29868c4c9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 127.207s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 784.761352] env[63531]: DEBUG oslo_concurrency.lockutils [req-421ffca7-f1d8-433b-9ca7-5d0c52b68cd0 req-1e64e7a3-e4ab-49e1-abb5-9d24b20aa328 service nova] Releasing lock "refresh_cache-1e5e28ef-38ce-47be-ac36-8bdf17ab0a47" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.796632] env[63531]: DEBUG nova.network.neutron [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 784.811039] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96d0974c-f47b-44e7-974c-104ce105014f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.822126] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e487202-0a3c-48f0-a422-f554d182a206 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.857115] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9925837e-b868-4801-a3b5-99a121361b44 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.865116] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8396dc74-7610-43ee-be44-33a58b46d993 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.879931] env[63531]: DEBUG nova.compute.provider_tree [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Updating inventory in ProviderTree for provider 9996a314-a62e-4282-8252-f387e7e3fabb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 784.920045] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117759, 'name': CreateVM_Task, 'duration_secs': 0.555037} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.920328] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 784.920978] env[63531]: DEBUG oslo_concurrency.lockutils [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 784.921210] env[63531]: DEBUG oslo_concurrency.lockutils [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 784.921542] env[63531]: DEBUG oslo_concurrency.lockutils [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 784.921797] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-45b16342-521b-4df1-a463-aea47c5a1015 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.926789] env[63531]: DEBUG oslo_vmware.api [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Waiting for the task: (returnval){ [ 784.926789] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]529b441d-a03a-8704-2786-f75592fc7a6f" [ 784.926789] env[63531]: _type = "Task" [ 784.926789] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.935622] env[63531]: DEBUG oslo_vmware.api [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]529b441d-a03a-8704-2786-f75592fc7a6f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.971437] env[63531]: DEBUG nova.network.neutron [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Updating instance_info_cache with network_info: [{"id": "e4133e0f-55c1-4b3f-847d-899dd1167eec", "address": "fa:16:3e:90:40:9b", "network": {"id": "c4f4c994-36ab-42ab-94f8-78f29cbafa50", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-994507698-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "57e9bf4ee0f1463fa67b0f8deff30fdc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "767a3a48-41d4-4a0c-961d-0024837f63bd", "external-id": "nsx-vlan-transportzone-141", "segmentation_id": 141, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4133e0f-55", "ovs_interfaceid": "e4133e0f-55c1-4b3f-847d-899dd1167eec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.008500] env[63531]: DEBUG oslo_vmware.api [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Task: {'id': task-1117760, 'name': Rename_Task, 'duration_secs': 0.223791} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.008638] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 785.008908] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0ccc6aa2-18d8-4535-b734-3e504b9431d8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.015425] env[63531]: DEBUG oslo_vmware.api [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Waiting for the task: (returnval){ [ 785.015425] env[63531]: value = "task-1117761" [ 785.015425] env[63531]: _type = "Task" [ 785.015425] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.023437] env[63531]: DEBUG oslo_vmware.api [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Task: {'id': task-1117761, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.037489] env[63531]: DEBUG nova.compute.manager [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 785.183043] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9e77b61d-e87e-4223-8227-0c685c107554 tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Acquiring lock "410ddf7a-e45b-4769-a3db-5363ce2096a7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.183359] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9e77b61d-e87e-4223-8227-0c685c107554 tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Lock "410ddf7a-e45b-4769-a3db-5363ce2096a7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.183596] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9e77b61d-e87e-4223-8227-0c685c107554 tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Acquiring lock "410ddf7a-e45b-4769-a3db-5363ce2096a7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.183781] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9e77b61d-e87e-4223-8227-0c685c107554 tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Lock "410ddf7a-e45b-4769-a3db-5363ce2096a7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.184093] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9e77b61d-e87e-4223-8227-0c685c107554 tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Lock "410ddf7a-e45b-4769-a3db-5363ce2096a7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.186540] env[63531]: INFO nova.compute.manager [None req-9e77b61d-e87e-4223-8227-0c685c107554 tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Terminating instance [ 785.189833] env[63531]: DEBUG nova.compute.manager [None req-9e77b61d-e87e-4223-8227-0c685c107554 tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 785.190182] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-abf9080a-bcb5-461d-a9f3-410df4966f49 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.199221] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-476dd7f9-6eff-4672-b87b-731b7584d0be {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.229498] env[63531]: WARNING nova.virt.vmwareapi.driver [None req-9e77b61d-e87e-4223-8227-0c685c107554 tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 410ddf7a-e45b-4769-a3db-5363ce2096a7 could not be found. [ 785.229770] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-9e77b61d-e87e-4223-8227-0c685c107554 tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 785.230145] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1d97b14a-6166-4afb-bbb6-2893c6b05d31 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.238444] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18119bb7-f7bb-45de-912f-809b9b7e8b44 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.268900] env[63531]: WARNING nova.virt.vmwareapi.vmops [None req-9e77b61d-e87e-4223-8227-0c685c107554 tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 410ddf7a-e45b-4769-a3db-5363ce2096a7 could not be found. [ 785.269245] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-9e77b61d-e87e-4223-8227-0c685c107554 tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 785.269495] env[63531]: INFO nova.compute.manager [None req-9e77b61d-e87e-4223-8227-0c685c107554 tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Took 0.08 seconds to destroy the instance on the hypervisor. [ 785.269819] env[63531]: DEBUG oslo.service.loopingcall [None req-9e77b61d-e87e-4223-8227-0c685c107554 tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 785.270161] env[63531]: DEBUG nova.compute.manager [-] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 785.270376] env[63531]: DEBUG nova.network.neutron [-] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 785.401325] env[63531]: ERROR nova.scheduler.client.report [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [req-e912478d-eb83-42e5-9e1c-c44b5f64e89b] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 9996a314-a62e-4282-8252-f387e7e3fabb. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-e912478d-eb83-42e5-9e1c-c44b5f64e89b"}]} [ 785.419020] env[63531]: DEBUG nova.scheduler.client.report [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Refreshing inventories for resource provider 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 785.447406] env[63531]: DEBUG oslo_vmware.api [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]529b441d-a03a-8704-2786-f75592fc7a6f, 'name': SearchDatastore_Task, 'duration_secs': 0.012495} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.449988] env[63531]: DEBUG nova.scheduler.client.report [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Updating ProviderTree inventory for provider 9996a314-a62e-4282-8252-f387e7e3fabb from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 785.450280] env[63531]: DEBUG nova.compute.provider_tree [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Updating inventory in ProviderTree for provider 9996a314-a62e-4282-8252-f387e7e3fabb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 785.452376] env[63531]: DEBUG oslo_concurrency.lockutils [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 785.452615] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 785.452852] env[63531]: DEBUG oslo_concurrency.lockutils [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 785.452999] env[63531]: DEBUG oslo_concurrency.lockutils [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 785.454278] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 785.455572] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0f3385a1-5cd7-47d6-914c-12dc90a751e0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.470258] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 785.470470] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 785.471887] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39fdcc6b-6c0e-4afc-8e15-2671c639a938 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.475637] env[63531]: DEBUG nova.scheduler.client.report [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Refreshing aggregate associations for resource provider 9996a314-a62e-4282-8252-f387e7e3fabb, aggregates: None {{(pid=63531) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 785.478333] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Releasing lock "refresh_cache-9a3fad46-a15a-451c-bdab-a3c8cc8add07" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 785.478651] env[63531]: DEBUG nova.compute.manager [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Instance network_info: |[{"id": "e4133e0f-55c1-4b3f-847d-899dd1167eec", "address": "fa:16:3e:90:40:9b", "network": {"id": "c4f4c994-36ab-42ab-94f8-78f29cbafa50", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-994507698-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "57e9bf4ee0f1463fa67b0f8deff30fdc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "767a3a48-41d4-4a0c-961d-0024837f63bd", "external-id": "nsx-vlan-transportzone-141", "segmentation_id": 141, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4133e0f-55", "ovs_interfaceid": "e4133e0f-55c1-4b3f-847d-899dd1167eec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 785.479460] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:90:40:9b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '767a3a48-41d4-4a0c-961d-0024837f63bd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e4133e0f-55c1-4b3f-847d-899dd1167eec', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 785.488641] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Creating folder: Project (57e9bf4ee0f1463fa67b0f8deff30fdc). Parent ref: group-v244585. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 785.490395] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-10aaf120-4811-4048-9482-6ac8c87a0f22 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.495400] env[63531]: DEBUG oslo_vmware.api [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Waiting for the task: (returnval){ [ 785.495400] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52c1bf05-9cff-05a0-b730-04f81073c99b" [ 785.495400] env[63531]: _type = "Task" [ 785.495400] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.504617] env[63531]: DEBUG oslo_vmware.api [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52c1bf05-9cff-05a0-b730-04f81073c99b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.507152] env[63531]: DEBUG nova.scheduler.client.report [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Refreshing trait associations for resource provider 9996a314-a62e-4282-8252-f387e7e3fabb, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE {{(pid=63531) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 785.509261] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Created folder: Project (57e9bf4ee0f1463fa67b0f8deff30fdc) in parent group-v244585. [ 785.510329] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Creating folder: Instances. Parent ref: group-v244662. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 785.510329] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-96a375fd-4bcc-40b1-bd52-a409a50e9635 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.521180] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Created folder: Instances in parent group-v244662. [ 785.521426] env[63531]: DEBUG oslo.service.loopingcall [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 785.521997] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 785.522543] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e8d3af85-d9a5-4b89-848e-5db66f4a62a3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.545582] env[63531]: DEBUG nova.compute.manager [None req-1171cdb2-37ed-4637-9a88-be822ac0cef5 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 785.546637] env[63531]: DEBUG oslo_vmware.api [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Task: {'id': task-1117761, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.550061] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46257c02-cb75-42c8-acd8-4c8fabd771b0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.565807] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 785.565807] env[63531]: value = "task-1117764" [ 785.565807] env[63531]: _type = "Task" [ 785.565807] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.573290] env[63531]: DEBUG oslo_concurrency.lockutils [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.575539] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117764, 'name': CreateVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.686105] env[63531]: DEBUG nova.compute.manager [req-8e985fe6-013a-4ffb-a31b-128b3ca5e7ca req-860d820d-e552-4108-8513-5f7e3d7e41ab service nova] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Received event network-changed-e4133e0f-55c1-4b3f-847d-899dd1167eec {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 785.686469] env[63531]: DEBUG nova.compute.manager [req-8e985fe6-013a-4ffb-a31b-128b3ca5e7ca req-860d820d-e552-4108-8513-5f7e3d7e41ab service nova] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Refreshing instance network info cache due to event network-changed-e4133e0f-55c1-4b3f-847d-899dd1167eec. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 785.686839] env[63531]: DEBUG oslo_concurrency.lockutils [req-8e985fe6-013a-4ffb-a31b-128b3ca5e7ca req-860d820d-e552-4108-8513-5f7e3d7e41ab service nova] Acquiring lock "refresh_cache-9a3fad46-a15a-451c-bdab-a3c8cc8add07" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 785.687167] env[63531]: DEBUG oslo_concurrency.lockutils [req-8e985fe6-013a-4ffb-a31b-128b3ca5e7ca req-860d820d-e552-4108-8513-5f7e3d7e41ab service nova] Acquired lock "refresh_cache-9a3fad46-a15a-451c-bdab-a3c8cc8add07" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 785.687451] env[63531]: DEBUG nova.network.neutron [req-8e985fe6-013a-4ffb-a31b-128b3ca5e7ca req-860d820d-e552-4108-8513-5f7e3d7e41ab service nova] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Refreshing network info cache for port e4133e0f-55c1-4b3f-847d-899dd1167eec {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 785.958488] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-140f7822-597b-422a-82d8-b40695854779 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.966508] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bef2155e-de02-4248-b0c6-9124ce09726e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.001506] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5cc3d9a-9450-4720-bf21-998da54d18e0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.012094] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f5aa0c6-bff0-4700-9983-3a17996ece63 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.015976] env[63531]: DEBUG oslo_vmware.api [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52c1bf05-9cff-05a0-b730-04f81073c99b, 'name': SearchDatastore_Task, 'duration_secs': 0.023603} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.017261] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f07d48cb-472d-4dcd-92ac-8090d193e965 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.030111] env[63531]: DEBUG nova.compute.provider_tree [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Updating inventory in ProviderTree for provider 9996a314-a62e-4282-8252-f387e7e3fabb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 786.039495] env[63531]: DEBUG oslo_vmware.api [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Task: {'id': task-1117761, 'name': PowerOnVM_Task} progress is 92%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.039644] env[63531]: DEBUG oslo_vmware.api [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Waiting for the task: (returnval){ [ 786.039644] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]524a930b-7043-bfed-ed87-f95171e0694c" [ 786.039644] env[63531]: _type = "Task" [ 786.039644] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.048849] env[63531]: DEBUG oslo_vmware.api [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]524a930b-7043-bfed-ed87-f95171e0694c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.067148] env[63531]: INFO nova.compute.manager [None req-1171cdb2-37ed-4637-9a88-be822ac0cef5 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] instance snapshotting [ 786.073685] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b5fae7e-f7d5-453d-b232-7a8f2145d357 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.082172] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117764, 'name': CreateVM_Task, 'duration_secs': 0.426748} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.096504] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 786.097425] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 786.097590] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.097932] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 786.098673] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e03f8cd1-c2cc-4bc0-a4d2-1e6a852c535f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.101252] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-78ca9e9f-e830-4232-9d52-6bc31ab51c16 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.106844] env[63531]: DEBUG oslo_vmware.api [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 786.106844] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5235c703-7f3f-c0dd-1b06-f2f137de5b05" [ 786.106844] env[63531]: _type = "Task" [ 786.106844] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.120611] env[63531]: DEBUG oslo_vmware.api [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5235c703-7f3f-c0dd-1b06-f2f137de5b05, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.121548] env[63531]: DEBUG nova.network.neutron [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Port f178c862-e753-40d4-b7cd-3eabe7eda39e binding to destination host cpu-1 is already ACTIVE {{(pid=63531) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 786.121842] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Acquiring lock "refresh_cache-5e9042b3-4dc2-4fa3-a664-c4b49a22e400" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 786.122110] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Acquired lock "refresh_cache-5e9042b3-4dc2-4fa3-a664-c4b49a22e400" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.122378] env[63531]: DEBUG nova.network.neutron [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 786.528188] env[63531]: DEBUG oslo_vmware.api [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Task: {'id': task-1117761, 'name': PowerOnVM_Task, 'duration_secs': 1.155615} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.528464] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 786.528669] env[63531]: INFO nova.compute.manager [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Took 8.31 seconds to spawn the instance on the hypervisor. [ 786.528861] env[63531]: DEBUG nova.compute.manager [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 786.529882] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6ac7f6f-b82c-40c9-b186-c4808da3eb6d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.548269] env[63531]: DEBUG nova.network.neutron [req-8e985fe6-013a-4ffb-a31b-128b3ca5e7ca req-860d820d-e552-4108-8513-5f7e3d7e41ab service nova] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Updated VIF entry in instance network info cache for port e4133e0f-55c1-4b3f-847d-899dd1167eec. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 786.548594] env[63531]: DEBUG nova.network.neutron [req-8e985fe6-013a-4ffb-a31b-128b3ca5e7ca req-860d820d-e552-4108-8513-5f7e3d7e41ab service nova] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Updating instance_info_cache with network_info: [{"id": "e4133e0f-55c1-4b3f-847d-899dd1167eec", "address": "fa:16:3e:90:40:9b", "network": {"id": "c4f4c994-36ab-42ab-94f8-78f29cbafa50", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-994507698-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "57e9bf4ee0f1463fa67b0f8deff30fdc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "767a3a48-41d4-4a0c-961d-0024837f63bd", "external-id": "nsx-vlan-transportzone-141", "segmentation_id": 141, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4133e0f-55", "ovs_interfaceid": "e4133e0f-55c1-4b3f-847d-899dd1167eec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 786.568075] env[63531]: DEBUG oslo_vmware.api [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]524a930b-7043-bfed-ed87-f95171e0694c, 'name': SearchDatastore_Task, 'duration_secs': 0.013434} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.569112] env[63531]: DEBUG oslo_concurrency.lockutils [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 786.569405] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47/1e5e28ef-38ce-47be-ac36-8bdf17ab0a47.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 786.569666] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5317a1dc-e94e-42f1-b926-b57658f38fb4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.576217] env[63531]: DEBUG oslo_vmware.api [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Waiting for the task: (returnval){ [ 786.576217] env[63531]: value = "task-1117765" [ 786.576217] env[63531]: _type = "Task" [ 786.576217] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.584518] env[63531]: DEBUG oslo_vmware.api [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1117765, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.585672] env[63531]: DEBUG nova.scheduler.client.report [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Updated inventory for provider 9996a314-a62e-4282-8252-f387e7e3fabb with generation 75 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 786.585919] env[63531]: DEBUG nova.compute.provider_tree [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Updating resource provider 9996a314-a62e-4282-8252-f387e7e3fabb generation from 75 to 76 during operation: update_inventory {{(pid=63531) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 786.586125] env[63531]: DEBUG nova.compute.provider_tree [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Updating inventory in ProviderTree for provider 9996a314-a62e-4282-8252-f387e7e3fabb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 786.613531] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-1171cdb2-37ed-4637-9a88-be822ac0cef5 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Creating Snapshot of the VM instance {{(pid=63531) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 786.614314] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-8436e170-0599-4e11-91a5-8084b533b0ff {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.623648] env[63531]: DEBUG oslo_vmware.api [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5235c703-7f3f-c0dd-1b06-f2f137de5b05, 'name': SearchDatastore_Task, 'duration_secs': 0.015169} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.625382] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 786.625669] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 786.628123] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 786.628123] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.628123] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 786.629264] env[63531]: DEBUG oslo_vmware.api [None req-1171cdb2-37ed-4637-9a88-be822ac0cef5 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Waiting for the task: (returnval){ [ 786.629264] env[63531]: value = "task-1117766" [ 786.629264] env[63531]: _type = "Task" [ 786.629264] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.629500] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3346a01b-6069-4759-bdea-785eea52f942 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.640548] env[63531]: DEBUG oslo_vmware.api [None req-1171cdb2-37ed-4637-9a88-be822ac0cef5 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117766, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.650408] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 786.650640] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 786.651480] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-adb07ad5-fbc9-484f-99b2-c0312fa87ac6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.658269] env[63531]: DEBUG oslo_vmware.api [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 786.658269] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52778bd4-84ca-0d38-6f34-69bc62f6a65c" [ 786.658269] env[63531]: _type = "Task" [ 786.658269] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.666751] env[63531]: DEBUG oslo_vmware.api [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52778bd4-84ca-0d38-6f34-69bc62f6a65c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.672159] env[63531]: DEBUG nova.network.neutron [-] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 787.059161] env[63531]: DEBUG oslo_concurrency.lockutils [req-8e985fe6-013a-4ffb-a31b-128b3ca5e7ca req-860d820d-e552-4108-8513-5f7e3d7e41ab service nova] Releasing lock "refresh_cache-9a3fad46-a15a-451c-bdab-a3c8cc8add07" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 787.070747] env[63531]: INFO nova.compute.manager [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Took 40.33 seconds to build instance. [ 787.087926] env[63531]: DEBUG oslo_vmware.api [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1117765, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.091899] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.701s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.092437] env[63531]: DEBUG nova.compute.manager [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 787.096095] env[63531]: DEBUG oslo_concurrency.lockutils [None req-785bdba5-357d-4d51-af3f-e636dea49e89 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.694s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 787.096450] env[63531]: DEBUG nova.objects.instance [None req-785bdba5-357d-4d51-af3f-e636dea49e89 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Lazy-loading 'resources' on Instance uuid c3920e60-b27f-496b-b616-afb23ab0bc26 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 787.145853] env[63531]: DEBUG oslo_vmware.api [None req-1171cdb2-37ed-4637-9a88-be822ac0cef5 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117766, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.170538] env[63531]: DEBUG oslo_vmware.api [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52778bd4-84ca-0d38-6f34-69bc62f6a65c, 'name': SearchDatastore_Task, 'duration_secs': 0.011203} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.172108] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-259ed2cd-9a3c-4a6f-aec3-7fb8b48c4e84 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.178968] env[63531]: INFO nova.compute.manager [-] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Took 1.91 seconds to deallocate network for instance. [ 787.186039] env[63531]: DEBUG oslo_vmware.api [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 787.186039] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5288f3c7-9f17-d77d-ea65-410523758afa" [ 787.186039] env[63531]: _type = "Task" [ 787.186039] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.201729] env[63531]: DEBUG oslo_vmware.api [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5288f3c7-9f17-d77d-ea65-410523758afa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.229527] env[63531]: DEBUG nova.network.neutron [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Updating instance_info_cache with network_info: [{"id": "f178c862-e753-40d4-b7cd-3eabe7eda39e", "address": "fa:16:3e:a6:d3:d2", "network": {"id": "5439232a-ec6d-4184-a9f1-f7be5c1526bc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.35", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "54fbef8a6ee0488f820706d2036e9c99", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f267bcdd-0daa-4337-9709-5fc060c267d8", "external-id": "nsx-vlan-transportzone-308", "segmentation_id": 308, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf178c862-e7", "ovs_interfaceid": "f178c862-e753-40d4-b7cd-3eabe7eda39e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 787.573056] env[63531]: DEBUG oslo_concurrency.lockutils [None req-948a832c-14fa-4a55-8b1b-41e698d490db tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Lock "0197e03d-6c36-4e73-9472-8e3319eb89ce" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 104.507s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.587831] env[63531]: DEBUG oslo_vmware.api [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1117765, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.785458} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.587965] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47/1e5e28ef-38ce-47be-ac36-8bdf17ab0a47.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 787.588463] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 787.588614] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-531bc23b-e0f7-4694-9cb2-3bd80c7396bb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.595672] env[63531]: DEBUG oslo_vmware.api [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Waiting for the task: (returnval){ [ 787.595672] env[63531]: value = "task-1117767" [ 787.595672] env[63531]: _type = "Task" [ 787.595672] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.598886] env[63531]: DEBUG nova.compute.utils [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 787.602162] env[63531]: DEBUG nova.compute.manager [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 787.602288] env[63531]: DEBUG nova.network.neutron [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 787.618525] env[63531]: DEBUG oslo_vmware.api [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1117767, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.648726] env[63531]: DEBUG oslo_vmware.api [None req-1171cdb2-37ed-4637-9a88-be822ac0cef5 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117766, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.667794] env[63531]: DEBUG nova.policy [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1ea586d8e4f947d2902c130530d42466', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f9056486e3b640d0bcdb9369ed5d67ad', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 787.697758] env[63531]: DEBUG oslo_vmware.api [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5288f3c7-9f17-d77d-ea65-410523758afa, 'name': SearchDatastore_Task, 'duration_secs': 0.060675} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.698061] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 787.698402] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 9a3fad46-a15a-451c-bdab-a3c8cc8add07/9a3fad46-a15a-451c-bdab-a3c8cc8add07.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 787.698730] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-aed049f0-0df2-4254-96f9-311397cb30fd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.705172] env[63531]: DEBUG oslo_vmware.api [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 787.705172] env[63531]: value = "task-1117768" [ 787.705172] env[63531]: _type = "Task" [ 787.705172] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.714091] env[63531]: DEBUG oslo_vmware.api [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1117768, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.730807] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Releasing lock "refresh_cache-5e9042b3-4dc2-4fa3-a664-c4b49a22e400" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 787.751524] env[63531]: INFO nova.compute.manager [None req-9e77b61d-e87e-4223-8227-0c685c107554 tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Took 0.57 seconds to detach 1 volumes for instance. [ 787.754897] env[63531]: DEBUG nova.compute.manager [None req-9e77b61d-e87e-4223-8227-0c685c107554 tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Deleting volume: 2a24c282-f681-40bf-a19c-5c711d342687 {{(pid=63531) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 788.011343] env[63531]: DEBUG nova.compute.manager [req-2af3cd5e-121d-4dd9-9536-00a2f8e047ec req-d17809fc-ac61-4048-b041-35aafad8bb68 service nova] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Received event network-vif-deleted-c173088d-cfd8-4f05-8936-1d0d5927ae34 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 788.038952] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-589d4a92-269d-41e6-8d37-3034dff0e87c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.046898] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64adf366-958a-4b84-b213-fca5da3f9f06 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.082577] env[63531]: DEBUG nova.compute.manager [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] [instance: 43c29443-1210-4ee2-95c7-1257de308287] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 788.086129] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d169181e-e5d7-4791-afaa-dbf6af8f9d4d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.094542] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28b3796a-0830-40bd-bd0b-dc18649fac2c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.102660] env[63531]: DEBUG nova.compute.manager [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 788.114117] env[63531]: DEBUG nova.network.neutron [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Successfully created port: 44f13e67-ffdf-401f-bdfe-fbd57e4f4d63 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 788.117570] env[63531]: DEBUG nova.compute.provider_tree [None req-785bdba5-357d-4d51-af3f-e636dea49e89 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 788.122197] env[63531]: DEBUG oslo_vmware.api [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1117767, 'name': ExtendVirtualDisk_Task} progress is 50%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.145477] env[63531]: DEBUG oslo_vmware.api [None req-1171cdb2-37ed-4637-9a88-be822ac0cef5 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117766, 'name': CreateSnapshot_Task, 'duration_secs': 1.251765} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.146320] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-1171cdb2-37ed-4637-9a88-be822ac0cef5 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Created Snapshot of the VM instance {{(pid=63531) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 788.147077] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-991c8679-0551-489f-8ec1-5b9425a064d2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.216757] env[63531]: DEBUG oslo_vmware.api [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1117768, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.236429] env[63531]: DEBUG nova.compute.manager [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=63531) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 788.236650] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 788.314461] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9e77b61d-e87e-4223-8227-0c685c107554 tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 788.611798] env[63531]: DEBUG oslo_vmware.api [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1117767, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.532959} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.612443] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 788.613572] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfd20ad9-0012-41bf-8e81-c0127dcbc31a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.620485] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 788.626592] env[63531]: DEBUG nova.scheduler.client.report [None req-785bdba5-357d-4d51-af3f-e636dea49e89 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 788.658905] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Reconfiguring VM instance instance-00000035 to attach disk [datastore1] 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47/1e5e28ef-38ce-47be-ac36-8bdf17ab0a47.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 788.660729] env[63531]: DEBUG oslo_concurrency.lockutils [None req-785bdba5-357d-4d51-af3f-e636dea49e89 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.565s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 788.664377] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c1b69000-d347-46c1-9716-d07af3e4ea82 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.688809] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0e74d968-ab93-4f14-92a6-24657e7c5585 tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.008s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 788.688809] env[63531]: DEBUG nova.objects.instance [None req-0e74d968-ab93-4f14-92a6-24657e7c5585 tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Lazy-loading 'resources' on Instance uuid e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 788.697564] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-1171cdb2-37ed-4637-9a88-be822ac0cef5 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Creating linked-clone VM from snapshot {{(pid=63531) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 788.700574] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-1e482b4f-5ae0-4ddb-b9a4-2fcf1437af74 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.716228] env[63531]: DEBUG oslo_vmware.api [None req-1171cdb2-37ed-4637-9a88-be822ac0cef5 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Waiting for the task: (returnval){ [ 788.716228] env[63531]: value = "task-1117771" [ 788.716228] env[63531]: _type = "Task" [ 788.716228] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.716516] env[63531]: DEBUG oslo_vmware.api [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Waiting for the task: (returnval){ [ 788.716516] env[63531]: value = "task-1117770" [ 788.716516] env[63531]: _type = "Task" [ 788.716516] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.733141] env[63531]: DEBUG oslo_vmware.api [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1117768, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.989856} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.734249] env[63531]: INFO nova.scheduler.client.report [None req-785bdba5-357d-4d51-af3f-e636dea49e89 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Deleted allocations for instance c3920e60-b27f-496b-b616-afb23ab0bc26 [ 788.738247] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 9a3fad46-a15a-451c-bdab-a3c8cc8add07/9a3fad46-a15a-451c-bdab-a3c8cc8add07.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 788.738247] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 788.748497] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-14deffea-28ef-458a-99ae-0cb43fcfecb9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.751393] env[63531]: DEBUG oslo_vmware.api [None req-1171cdb2-37ed-4637-9a88-be822ac0cef5 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117771, 'name': CloneVM_Task} progress is 11%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.752023] env[63531]: DEBUG oslo_vmware.api [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1117770, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.757496] env[63531]: DEBUG oslo_vmware.api [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 788.757496] env[63531]: value = "task-1117772" [ 788.757496] env[63531]: _type = "Task" [ 788.757496] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.769915] env[63531]: DEBUG oslo_vmware.api [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1117772, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.072411] env[63531]: DEBUG nova.compute.manager [req-2d172f9c-e4b5-4b83-9212-86c332133016 req-8bb0f07d-603a-45ee-a57b-7daca30872dd service nova] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Received event network-changed-b4141cf4-dab5-4033-8cc5-06966340323b {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 789.072600] env[63531]: DEBUG nova.compute.manager [req-2d172f9c-e4b5-4b83-9212-86c332133016 req-8bb0f07d-603a-45ee-a57b-7daca30872dd service nova] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Refreshing instance network info cache due to event network-changed-b4141cf4-dab5-4033-8cc5-06966340323b. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 789.072816] env[63531]: DEBUG oslo_concurrency.lockutils [req-2d172f9c-e4b5-4b83-9212-86c332133016 req-8bb0f07d-603a-45ee-a57b-7daca30872dd service nova] Acquiring lock "refresh_cache-0197e03d-6c36-4e73-9472-8e3319eb89ce" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 789.072955] env[63531]: DEBUG oslo_concurrency.lockutils [req-2d172f9c-e4b5-4b83-9212-86c332133016 req-8bb0f07d-603a-45ee-a57b-7daca30872dd service nova] Acquired lock "refresh_cache-0197e03d-6c36-4e73-9472-8e3319eb89ce" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.073476] env[63531]: DEBUG nova.network.neutron [req-2d172f9c-e4b5-4b83-9212-86c332133016 req-8bb0f07d-603a-45ee-a57b-7daca30872dd service nova] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Refreshing network info cache for port b4141cf4-dab5-4033-8cc5-06966340323b {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 789.134778] env[63531]: DEBUG nova.compute.manager [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 789.166893] env[63531]: DEBUG nova.virt.hardware [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 789.167315] env[63531]: DEBUG nova.virt.hardware [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 789.170134] env[63531]: DEBUG nova.virt.hardware [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 789.170341] env[63531]: DEBUG nova.virt.hardware [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 789.170631] env[63531]: DEBUG nova.virt.hardware [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 789.170716] env[63531]: DEBUG nova.virt.hardware [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 789.170875] env[63531]: DEBUG nova.virt.hardware [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 789.171067] env[63531]: DEBUG nova.virt.hardware [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 789.171242] env[63531]: DEBUG nova.virt.hardware [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 789.171407] env[63531]: DEBUG nova.virt.hardware [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 789.171578] env[63531]: DEBUG nova.virt.hardware [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 789.172753] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8915847d-fcc1-45a6-b46e-b654cf00214b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.184425] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-300866b1-1de4-4db4-9f86-9cd814f65915 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.237385] env[63531]: DEBUG oslo_vmware.api [None req-1171cdb2-37ed-4637-9a88-be822ac0cef5 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117771, 'name': CloneVM_Task} progress is 94%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.238907] env[63531]: DEBUG oslo_vmware.api [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1117770, 'name': ReconfigVM_Task, 'duration_secs': 0.316623} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.242740] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Reconfigured VM instance instance-00000035 to attach disk [datastore1] 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47/1e5e28ef-38ce-47be-ac36-8bdf17ab0a47.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 789.245833] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c59b8293-1b50-49ee-85db-8ec3931f39a0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.252786] env[63531]: DEBUG oslo_vmware.api [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Waiting for the task: (returnval){ [ 789.252786] env[63531]: value = "task-1117773" [ 789.252786] env[63531]: _type = "Task" [ 789.252786] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.256243] env[63531]: DEBUG oslo_concurrency.lockutils [None req-785bdba5-357d-4d51-af3f-e636dea49e89 tempest-ServerRescueTestJSONUnderV235-68934667 tempest-ServerRescueTestJSONUnderV235-68934667-project-member] Lock "c3920e60-b27f-496b-b616-afb23ab0bc26" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.817s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 789.273209] env[63531]: DEBUG oslo_vmware.api [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1117773, 'name': Rename_Task} progress is 10%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.278047] env[63531]: DEBUG oslo_vmware.api [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1117772, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.145562} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.278252] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 789.279252] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd1b2ac7-4936-4ec9-8bdc-ff3cf9cdd2bb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.304424] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Reconfiguring VM instance instance-00000036 to attach disk [datastore1] 9a3fad46-a15a-451c-bdab-a3c8cc8add07/9a3fad46-a15a-451c-bdab-a3c8cc8add07.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 789.308259] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b887b0c8-4429-40a3-8400-02903a0abcb1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.329237] env[63531]: DEBUG oslo_vmware.api [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 789.329237] env[63531]: value = "task-1117774" [ 789.329237] env[63531]: _type = "Task" [ 789.329237] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.345043] env[63531]: DEBUG oslo_vmware.api [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1117774, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.693683] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17ad8c65-5625-4f3d-a789-10c7951aec3b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.702112] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bff911d-8850-4db7-b0d9-8a9c679cb684 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.742519] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dfde3b4-6f27-48c7-bca8-c342ca7c0452 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.754229] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9570813c-9ca6-4dd3-9cd5-da7be9beef7b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.760118] env[63531]: DEBUG oslo_vmware.api [None req-1171cdb2-37ed-4637-9a88-be822ac0cef5 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117771, 'name': CloneVM_Task} progress is 94%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.769063] env[63531]: DEBUG oslo_vmware.api [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1117773, 'name': Rename_Task, 'duration_secs': 0.191693} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.777771] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 789.778453] env[63531]: DEBUG nova.compute.provider_tree [None req-0e74d968-ab93-4f14-92a6-24657e7c5585 tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 789.779822] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7ebba135-2eba-44fc-a56d-50c011b5f4fb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.788823] env[63531]: DEBUG oslo_vmware.api [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Waiting for the task: (returnval){ [ 789.788823] env[63531]: value = "task-1117775" [ 789.788823] env[63531]: _type = "Task" [ 789.788823] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.802987] env[63531]: DEBUG oslo_vmware.api [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1117775, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.840625] env[63531]: DEBUG oslo_vmware.api [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1117774, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.088149] env[63531]: DEBUG nova.network.neutron [req-2d172f9c-e4b5-4b83-9212-86c332133016 req-8bb0f07d-603a-45ee-a57b-7daca30872dd service nova] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Updated VIF entry in instance network info cache for port b4141cf4-dab5-4033-8cc5-06966340323b. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 790.088752] env[63531]: DEBUG nova.network.neutron [req-2d172f9c-e4b5-4b83-9212-86c332133016 req-8bb0f07d-603a-45ee-a57b-7daca30872dd service nova] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Updating instance_info_cache with network_info: [{"id": "b4141cf4-dab5-4033-8cc5-06966340323b", "address": "fa:16:3e:72:a7:59", "network": {"id": "7af8bbee-a43a-40b5-99c4-aee7e19b4093", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1560507257-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.195", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "68eaa455a44b4d96936fe8273629279a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ed82fc44-f216-40a7-a68b-62d76ffdb1f8", "external-id": "nsx-vlan-transportzone-236", "segmentation_id": 236, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4141cf4-da", "ovs_interfaceid": "b4141cf4-dab5-4033-8cc5-06966340323b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 790.167359] env[63531]: DEBUG nova.compute.manager [req-19cb059f-7396-483e-b581-6409aa670c14 req-04d25a5a-6830-4d8a-bff3-e5b45b9d5c2a service nova] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Received event network-vif-plugged-44f13e67-ffdf-401f-bdfe-fbd57e4f4d63 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 790.168112] env[63531]: DEBUG oslo_concurrency.lockutils [req-19cb059f-7396-483e-b581-6409aa670c14 req-04d25a5a-6830-4d8a-bff3-e5b45b9d5c2a service nova] Acquiring lock "48043b5d-c0d3-4e8a-8d28-5a408d800e11-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.168112] env[63531]: DEBUG oslo_concurrency.lockutils [req-19cb059f-7396-483e-b581-6409aa670c14 req-04d25a5a-6830-4d8a-bff3-e5b45b9d5c2a service nova] Lock "48043b5d-c0d3-4e8a-8d28-5a408d800e11-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 790.168624] env[63531]: DEBUG oslo_concurrency.lockutils [req-19cb059f-7396-483e-b581-6409aa670c14 req-04d25a5a-6830-4d8a-bff3-e5b45b9d5c2a service nova] Lock "48043b5d-c0d3-4e8a-8d28-5a408d800e11-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 790.169051] env[63531]: DEBUG nova.compute.manager [req-19cb059f-7396-483e-b581-6409aa670c14 req-04d25a5a-6830-4d8a-bff3-e5b45b9d5c2a service nova] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] No waiting events found dispatching network-vif-plugged-44f13e67-ffdf-401f-bdfe-fbd57e4f4d63 {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 790.169491] env[63531]: WARNING nova.compute.manager [req-19cb059f-7396-483e-b581-6409aa670c14 req-04d25a5a-6830-4d8a-bff3-e5b45b9d5c2a service nova] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Received unexpected event network-vif-plugged-44f13e67-ffdf-401f-bdfe-fbd57e4f4d63 for instance with vm_state building and task_state spawning. [ 790.248582] env[63531]: DEBUG oslo_vmware.api [None req-1171cdb2-37ed-4637-9a88-be822ac0cef5 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117771, 'name': CloneVM_Task} progress is 94%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.283975] env[63531]: DEBUG nova.scheduler.client.report [None req-0e74d968-ab93-4f14-92a6-24657e7c5585 tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 790.300708] env[63531]: DEBUG oslo_vmware.api [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1117775, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.343957] env[63531]: DEBUG oslo_vmware.api [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1117774, 'name': ReconfigVM_Task, 'duration_secs': 0.690535} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.344290] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Reconfigured VM instance instance-00000036 to attach disk [datastore1] 9a3fad46-a15a-451c-bdab-a3c8cc8add07/9a3fad46-a15a-451c-bdab-a3c8cc8add07.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 790.344973] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e085e4e1-dee3-49be-96c9-330c0fb7d4ce {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.353924] env[63531]: DEBUG oslo_vmware.api [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 790.353924] env[63531]: value = "task-1117776" [ 790.353924] env[63531]: _type = "Task" [ 790.353924] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.365188] env[63531]: DEBUG oslo_vmware.api [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1117776, 'name': Rename_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.477140] env[63531]: DEBUG nova.network.neutron [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Successfully updated port: 44f13e67-ffdf-401f-bdfe-fbd57e4f4d63 {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 790.592175] env[63531]: DEBUG oslo_concurrency.lockutils [req-2d172f9c-e4b5-4b83-9212-86c332133016 req-8bb0f07d-603a-45ee-a57b-7daca30872dd service nova] Releasing lock "refresh_cache-0197e03d-6c36-4e73-9472-8e3319eb89ce" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 790.750863] env[63531]: DEBUG oslo_vmware.api [None req-1171cdb2-37ed-4637-9a88-be822ac0cef5 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117771, 'name': CloneVM_Task, 'duration_secs': 1.804756} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.751393] env[63531]: INFO nova.virt.vmwareapi.vmops [None req-1171cdb2-37ed-4637-9a88-be822ac0cef5 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Created linked-clone VM from snapshot [ 790.754199] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1020b4ce-9846-43cc-a313-904a0e740b03 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.767892] env[63531]: DEBUG nova.virt.vmwareapi.images [None req-1171cdb2-37ed-4637-9a88-be822ac0cef5 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Uploading image c172132b-5ac9-4ddb-885b-cb2c5fed2c85 {{(pid=63531) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 790.800275] env[63531]: DEBUG oslo_vmware.rw_handles [None req-1171cdb2-37ed-4637-9a88-be822ac0cef5 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 790.800275] env[63531]: value = "vm-244666" [ 790.800275] env[63531]: _type = "VirtualMachine" [ 790.800275] env[63531]: }. {{(pid=63531) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 790.801767] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0e74d968-ab93-4f14-92a6-24657e7c5585 tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.113s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 790.804683] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-1c9bb762-ac09-4cb6-b5e3-b11bbece6fa6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.808777] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 25.954s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 790.808976] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 790.809153] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63531) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 790.809450] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c52c5465-cb3f-44ae-a7bc-34ca246641a2 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.651s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 790.809661] env[63531]: DEBUG nova.objects.instance [None req-c52c5465-cb3f-44ae-a7bc-34ca246641a2 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Lazy-loading 'resources' on Instance uuid 645abb7a-cf73-4fb0-a9a0-49205f060d6f {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 790.812230] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d11c29a8-034d-42f1-a237-d6308ca8c3e6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.820315] env[63531]: DEBUG oslo_vmware.api [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1117775, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.823022] env[63531]: DEBUG oslo_vmware.rw_handles [None req-1171cdb2-37ed-4637-9a88-be822ac0cef5 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Lease: (returnval){ [ 790.823022] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52677688-3d34-b1e1-f53b-caaf97079783" [ 790.823022] env[63531]: _type = "HttpNfcLease" [ 790.823022] env[63531]: } obtained for exporting VM: (result){ [ 790.823022] env[63531]: value = "vm-244666" [ 790.823022] env[63531]: _type = "VirtualMachine" [ 790.823022] env[63531]: }. {{(pid=63531) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 790.823418] env[63531]: DEBUG oslo_vmware.api [None req-1171cdb2-37ed-4637-9a88-be822ac0cef5 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Waiting for the lease: (returnval){ [ 790.823418] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52677688-3d34-b1e1-f53b-caaf97079783" [ 790.823418] env[63531]: _type = "HttpNfcLease" [ 790.823418] env[63531]: } to be ready. {{(pid=63531) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 790.835598] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59123acd-f0e1-4ce9-9cda-ca16f62b19df {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.844470] env[63531]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 790.844470] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52677688-3d34-b1e1-f53b-caaf97079783" [ 790.844470] env[63531]: _type = "HttpNfcLease" [ 790.844470] env[63531]: } is initializing. {{(pid=63531) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 790.845608] env[63531]: INFO nova.scheduler.client.report [None req-0e74d968-ab93-4f14-92a6-24657e7c5585 tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Deleted allocations for instance e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74 [ 790.863047] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f3267ab-5713-450a-bb1f-44b73d02bbb5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.872579] env[63531]: DEBUG oslo_vmware.api [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1117776, 'name': Rename_Task, 'duration_secs': 0.218949} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.875747] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 790.878437] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c44e7ef2-0c36-4a1e-bb91-6cb630963d89 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.878693] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e81ad8f1-4e75-4144-96ff-29e76d7b3fe1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.910636] env[63531]: DEBUG oslo_vmware.api [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 790.910636] env[63531]: value = "task-1117778" [ 790.910636] env[63531]: _type = "Task" [ 790.910636] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.911067] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181250MB free_disk=169GB free_vcpus=48 pci_devices=None {{(pid=63531) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 790.912416] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.924505] env[63531]: DEBUG oslo_vmware.api [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1117778, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.984262] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquiring lock "refresh_cache-48043b5d-c0d3-4e8a-8d28-5a408d800e11" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.984262] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquired lock "refresh_cache-48043b5d-c0d3-4e8a-8d28-5a408d800e11" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.984262] env[63531]: DEBUG nova.network.neutron [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 791.301941] env[63531]: DEBUG oslo_vmware.api [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1117775, 'name': PowerOnVM_Task, 'duration_secs': 1.138773} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.301941] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 791.301941] env[63531]: INFO nova.compute.manager [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Took 10.55 seconds to spawn the instance on the hypervisor. [ 791.301941] env[63531]: DEBUG nova.compute.manager [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 791.302610] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae23974a-6935-48dd-979b-ba38406c09bb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.335707] env[63531]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 791.335707] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52677688-3d34-b1e1-f53b-caaf97079783" [ 791.335707] env[63531]: _type = "HttpNfcLease" [ 791.335707] env[63531]: } is ready. {{(pid=63531) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 791.336800] env[63531]: DEBUG oslo_vmware.rw_handles [None req-1171cdb2-37ed-4637-9a88-be822ac0cef5 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 791.336800] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52677688-3d34-b1e1-f53b-caaf97079783" [ 791.336800] env[63531]: _type = "HttpNfcLease" [ 791.336800] env[63531]: }. {{(pid=63531) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 791.338879] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8851bb1b-7c17-406e-85b3-248c4a8e6973 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.355082] env[63531]: DEBUG oslo_vmware.rw_handles [None req-1171cdb2-37ed-4637-9a88-be822ac0cef5 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f30212-f5b0-861d-c7c3-797189e87aa9/disk-0.vmdk from lease info. {{(pid=63531) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 791.355343] env[63531]: DEBUG oslo_vmware.rw_handles [None req-1171cdb2-37ed-4637-9a88-be822ac0cef5 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f30212-f5b0-861d-c7c3-797189e87aa9/disk-0.vmdk for reading. {{(pid=63531) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 791.429366] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0e74d968-ab93-4f14-92a6-24657e7c5585 tempest-VolumesAdminNegativeTest-1089851889 tempest-VolumesAdminNegativeTest-1089851889-project-member] Lock "e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.838s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 791.437427] env[63531]: DEBUG oslo_vmware.api [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1117778, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.484827] env[63531]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-69aebe29-c061-453a-90fe-1fd0d5c2c68d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.572755] env[63531]: DEBUG nova.network.neutron [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 791.733228] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Acquiring lock "45b05c54-6865-47a1-ba93-90ad3e1ba07e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 791.733527] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Lock "45b05c54-6865-47a1-ba93-90ad3e1ba07e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 791.828913] env[63531]: INFO nova.compute.manager [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Took 42.73 seconds to build instance. [ 791.845170] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04fef9a2-2878-44d9-8433-5d1aa1f9da0d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.869646] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-331ffc8b-d0eb-4b7e-ac04-1cfc8d57805f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.912683] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1373ce5b-9d18-4568-9c63-e1fc17601b51 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.929090] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cea3cd41-31ae-4c2e-a242-390fa60a3ca8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.944582] env[63531]: DEBUG oslo_vmware.api [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1117778, 'name': PowerOnVM_Task, 'duration_secs': 0.901331} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.953326] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 791.953326] env[63531]: INFO nova.compute.manager [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Took 8.61 seconds to spawn the instance on the hypervisor. [ 791.953326] env[63531]: DEBUG nova.compute.manager [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 791.954025] env[63531]: DEBUG nova.compute.provider_tree [None req-c52c5465-cb3f-44ae-a7bc-34ca246641a2 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 791.955692] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44d739ee-73f6-431b-8a7e-6ce7c0811d7b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.976053] env[63531]: DEBUG nova.network.neutron [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Updating instance_info_cache with network_info: [{"id": "44f13e67-ffdf-401f-bdfe-fbd57e4f4d63", "address": "fa:16:3e:c5:55:ae", "network": {"id": "95ed2c7d-3d1a-4736-8e90-104002f5c40b", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-73549344-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9056486e3b640d0bcdb9369ed5d67ad", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04e15990-16e1-4cb2-b0f0-06c362e68c5e", "external-id": "nsx-vlan-transportzone-555", "segmentation_id": 555, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44f13e67-ff", "ovs_interfaceid": "44f13e67-ffdf-401f-bdfe-fbd57e4f4d63", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.237983] env[63531]: DEBUG nova.compute.manager [req-f78e5d41-5dba-48a1-926b-e57f36039eb2 req-23369262-eaf1-4bac-83f3-c827d203af90 service nova] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Received event network-changed-44f13e67-ffdf-401f-bdfe-fbd57e4f4d63 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 792.237983] env[63531]: DEBUG nova.compute.manager [req-f78e5d41-5dba-48a1-926b-e57f36039eb2 req-23369262-eaf1-4bac-83f3-c827d203af90 service nova] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Refreshing instance network info cache due to event network-changed-44f13e67-ffdf-401f-bdfe-fbd57e4f4d63. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 792.237983] env[63531]: DEBUG oslo_concurrency.lockutils [req-f78e5d41-5dba-48a1-926b-e57f36039eb2 req-23369262-eaf1-4bac-83f3-c827d203af90 service nova] Acquiring lock "refresh_cache-48043b5d-c0d3-4e8a-8d28-5a408d800e11" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 792.331319] env[63531]: DEBUG oslo_concurrency.lockutils [None req-755cf195-df08-4cb4-bb0a-cd97f0f927b7 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Lock "1e5e28ef-38ce-47be-ac36-8bdf17ab0a47" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 109.223s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.460111] env[63531]: DEBUG nova.scheduler.client.report [None req-c52c5465-cb3f-44ae-a7bc-34ca246641a2 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 792.480771] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Releasing lock "refresh_cache-48043b5d-c0d3-4e8a-8d28-5a408d800e11" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.480771] env[63531]: DEBUG nova.compute.manager [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Instance network_info: |[{"id": "44f13e67-ffdf-401f-bdfe-fbd57e4f4d63", "address": "fa:16:3e:c5:55:ae", "network": {"id": "95ed2c7d-3d1a-4736-8e90-104002f5c40b", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-73549344-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9056486e3b640d0bcdb9369ed5d67ad", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04e15990-16e1-4cb2-b0f0-06c362e68c5e", "external-id": "nsx-vlan-transportzone-555", "segmentation_id": 555, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44f13e67-ff", "ovs_interfaceid": "44f13e67-ffdf-401f-bdfe-fbd57e4f4d63", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 792.480771] env[63531]: INFO nova.compute.manager [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Took 41.38 seconds to build instance. [ 792.481655] env[63531]: DEBUG oslo_concurrency.lockutils [req-f78e5d41-5dba-48a1-926b-e57f36039eb2 req-23369262-eaf1-4bac-83f3-c827d203af90 service nova] Acquired lock "refresh_cache-48043b5d-c0d3-4e8a-8d28-5a408d800e11" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.483901] env[63531]: DEBUG nova.network.neutron [req-f78e5d41-5dba-48a1-926b-e57f36039eb2 req-23369262-eaf1-4bac-83f3-c827d203af90 service nova] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Refreshing network info cache for port 44f13e67-ffdf-401f-bdfe-fbd57e4f4d63 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 792.483901] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c5:55:ae', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '04e15990-16e1-4cb2-b0f0-06c362e68c5e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '44f13e67-ffdf-401f-bdfe-fbd57e4f4d63', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 792.492869] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Creating folder: Project (f9056486e3b640d0bcdb9369ed5d67ad). Parent ref: group-v244585. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 792.494747] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-303b385e-10eb-4a89-8416-5668237c0b58 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.509466] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Created folder: Project (f9056486e3b640d0bcdb9369ed5d67ad) in parent group-v244585. [ 792.509466] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Creating folder: Instances. Parent ref: group-v244667. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 792.509466] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-691b34c2-d77c-462a-80c8-63c2b512da46 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.521780] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Created folder: Instances in parent group-v244667. [ 792.522728] env[63531]: DEBUG oslo.service.loopingcall [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 792.522728] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 792.522982] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1e2dcc3c-2759-4d94-8cff-97cd750250b3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.547237] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 792.547237] env[63531]: value = "task-1117781" [ 792.547237] env[63531]: _type = "Task" [ 792.547237] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.559372] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117781, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.843311] env[63531]: DEBUG nova.compute.manager [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 792.967183] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c52c5465-cb3f-44ae-a7bc-34ca246641a2 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.157s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.973851] env[63531]: DEBUG oslo_concurrency.lockutils [None req-fdd04ada-cc1b-4e03-9008-da80c3e4e411 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.747s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 792.973851] env[63531]: DEBUG nova.objects.instance [None req-fdd04ada-cc1b-4e03-9008-da80c3e4e411 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Lazy-loading 'resources' on Instance uuid f39c8807-1e15-41cf-899b-a1fbe0695d58 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 792.996992] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1684a07b-ff9f-4f42-90e7-88458e6e8b6b tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "9a3fad46-a15a-451c-bdab-a3c8cc8add07" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 105.626s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.001208] env[63531]: INFO nova.scheduler.client.report [None req-c52c5465-cb3f-44ae-a7bc-34ca246641a2 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Deleted allocations for instance 645abb7a-cf73-4fb0-a9a0-49205f060d6f [ 793.063558] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117781, 'name': CreateVM_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.381024] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.516169] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c52c5465-cb3f-44ae-a7bc-34ca246641a2 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Lock "645abb7a-cf73-4fb0-a9a0-49205f060d6f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.405s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.519236] env[63531]: DEBUG nova.network.neutron [req-f78e5d41-5dba-48a1-926b-e57f36039eb2 req-23369262-eaf1-4bac-83f3-c827d203af90 service nova] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Updated VIF entry in instance network info cache for port 44f13e67-ffdf-401f-bdfe-fbd57e4f4d63. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 793.519639] env[63531]: DEBUG nova.network.neutron [req-f78e5d41-5dba-48a1-926b-e57f36039eb2 req-23369262-eaf1-4bac-83f3-c827d203af90 service nova] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Updating instance_info_cache with network_info: [{"id": "44f13e67-ffdf-401f-bdfe-fbd57e4f4d63", "address": "fa:16:3e:c5:55:ae", "network": {"id": "95ed2c7d-3d1a-4736-8e90-104002f5c40b", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-73549344-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9056486e3b640d0bcdb9369ed5d67ad", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04e15990-16e1-4cb2-b0f0-06c362e68c5e", "external-id": "nsx-vlan-transportzone-555", "segmentation_id": 555, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44f13e67-ff", "ovs_interfaceid": "44f13e67-ffdf-401f-bdfe-fbd57e4f4d63", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 793.560450] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117781, 'name': CreateVM_Task, 'duration_secs': 0.511732} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.560689] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 793.561465] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.561864] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.562010] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 793.564995] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b0d94acb-f01d-40eb-9d50-ed733c36462b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.573901] env[63531]: DEBUG oslo_vmware.api [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for the task: (returnval){ [ 793.573901] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52f26f9f-96f4-4a03-0912-1ee912e1491e" [ 793.573901] env[63531]: _type = "Task" [ 793.573901] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.586304] env[63531]: DEBUG oslo_vmware.api [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52f26f9f-96f4-4a03-0912-1ee912e1491e, 'name': SearchDatastore_Task, 'duration_secs': 0.010435} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.586468] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 793.586598] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 793.586876] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.587084] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.587326] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 793.590499] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ed6169cd-58ab-4558-8753-b8c0b5e6c170 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.601867] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 793.601867] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 793.601867] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-506a53b3-3e13-40af-baf3-709da339b8ad {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.608676] env[63531]: DEBUG oslo_vmware.api [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for the task: (returnval){ [ 793.608676] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52eec2b7-7c5f-d639-e800-1785c8f3c2d7" [ 793.608676] env[63531]: _type = "Task" [ 793.608676] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.624017] env[63531]: DEBUG oslo_vmware.api [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52eec2b7-7c5f-d639-e800-1785c8f3c2d7, 'name': SearchDatastore_Task, 'duration_secs': 0.009441} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.624428] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1a7587d6-776c-4cfb-807e-51e74acc02cd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.632915] env[63531]: DEBUG oslo_vmware.api [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for the task: (returnval){ [ 793.632915] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]523385b3-e956-5946-765b-71779e771e6b" [ 793.632915] env[63531]: _type = "Task" [ 793.632915] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.644367] env[63531]: DEBUG oslo_vmware.api [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]523385b3-e956-5946-765b-71779e771e6b, 'name': SearchDatastore_Task, 'duration_secs': 0.009815} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.644367] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 793.644367] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] 48043b5d-c0d3-4e8a-8d28-5a408d800e11/48043b5d-c0d3-4e8a-8d28-5a408d800e11.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 793.644367] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-106c1705-32b6-46b6-bbbc-a50fd32bcf2d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.653397] env[63531]: DEBUG oslo_vmware.api [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for the task: (returnval){ [ 793.653397] env[63531]: value = "task-1117782" [ 793.653397] env[63531]: _type = "Task" [ 793.653397] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.664023] env[63531]: DEBUG oslo_vmware.api [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1117782, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.835407] env[63531]: DEBUG nova.compute.manager [req-118de9d7-f22e-44f3-abcd-9d2280330154 req-32807e4e-4b97-4db3-b688-fa45357659e9 service nova] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Received event network-changed-e4133e0f-55c1-4b3f-847d-899dd1167eec {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 793.835665] env[63531]: DEBUG nova.compute.manager [req-118de9d7-f22e-44f3-abcd-9d2280330154 req-32807e4e-4b97-4db3-b688-fa45357659e9 service nova] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Refreshing instance network info cache due to event network-changed-e4133e0f-55c1-4b3f-847d-899dd1167eec. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 793.836202] env[63531]: DEBUG oslo_concurrency.lockutils [req-118de9d7-f22e-44f3-abcd-9d2280330154 req-32807e4e-4b97-4db3-b688-fa45357659e9 service nova] Acquiring lock "refresh_cache-9a3fad46-a15a-451c-bdab-a3c8cc8add07" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.836202] env[63531]: DEBUG oslo_concurrency.lockutils [req-118de9d7-f22e-44f3-abcd-9d2280330154 req-32807e4e-4b97-4db3-b688-fa45357659e9 service nova] Acquired lock "refresh_cache-9a3fad46-a15a-451c-bdab-a3c8cc8add07" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.837049] env[63531]: DEBUG nova.network.neutron [req-118de9d7-f22e-44f3-abcd-9d2280330154 req-32807e4e-4b97-4db3-b688-fa45357659e9 service nova] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Refreshing network info cache for port e4133e0f-55c1-4b3f-847d-899dd1167eec {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 793.922970] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12c1a13d-7d62-4241-8621-78ee8d3f9365 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.936468] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5173997-360a-4d77-b2b9-a2635a7e7657 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.978437] env[63531]: DEBUG nova.compute.manager [None req-0acd5bb8-82ed-4cf7-9ea3-96d802548d79 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 793.979631] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-000e475a-ffbe-4de5-81c5-9c2c25748c79 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.983907] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccb8d0f3-ce8e-4436-86d6-11d3cce94abc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.998112] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5f30086-9368-4b0c-99e4-1d7d87c5cf98 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.013444] env[63531]: DEBUG nova.compute.provider_tree [None req-fdd04ada-cc1b-4e03-9008-da80c3e4e411 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 794.023491] env[63531]: DEBUG oslo_concurrency.lockutils [req-f78e5d41-5dba-48a1-926b-e57f36039eb2 req-23369262-eaf1-4bac-83f3-c827d203af90 service nova] Releasing lock "refresh_cache-48043b5d-c0d3-4e8a-8d28-5a408d800e11" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.169981] env[63531]: DEBUG oslo_vmware.api [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1117782, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.48597} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.169981] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] 48043b5d-c0d3-4e8a-8d28-5a408d800e11/48043b5d-c0d3-4e8a-8d28-5a408d800e11.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 794.169981] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 794.171062] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9b778abf-ba5d-4862-b13c-25daf6f468b3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.177408] env[63531]: DEBUG oslo_vmware.api [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for the task: (returnval){ [ 794.177408] env[63531]: value = "task-1117783" [ 794.177408] env[63531]: _type = "Task" [ 794.177408] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.187891] env[63531]: DEBUG oslo_vmware.api [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1117783, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.501955] env[63531]: INFO nova.compute.manager [None req-0acd5bb8-82ed-4cf7-9ea3-96d802548d79 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] instance snapshotting [ 794.507670] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c73bbbaa-caa6-4d51-884c-ff538710c237 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.528548] env[63531]: DEBUG nova.scheduler.client.report [None req-fdd04ada-cc1b-4e03-9008-da80c3e4e411 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 794.532917] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ae80762-47be-463f-8e0c-ff7957c3b819 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.623392] env[63531]: DEBUG nova.network.neutron [req-118de9d7-f22e-44f3-abcd-9d2280330154 req-32807e4e-4b97-4db3-b688-fa45357659e9 service nova] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Updated VIF entry in instance network info cache for port e4133e0f-55c1-4b3f-847d-899dd1167eec. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 794.623722] env[63531]: DEBUG nova.network.neutron [req-118de9d7-f22e-44f3-abcd-9d2280330154 req-32807e4e-4b97-4db3-b688-fa45357659e9 service nova] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Updating instance_info_cache with network_info: [{"id": "e4133e0f-55c1-4b3f-847d-899dd1167eec", "address": "fa:16:3e:90:40:9b", "network": {"id": "c4f4c994-36ab-42ab-94f8-78f29cbafa50", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-994507698-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "57e9bf4ee0f1463fa67b0f8deff30fdc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "767a3a48-41d4-4a0c-961d-0024837f63bd", "external-id": "nsx-vlan-transportzone-141", "segmentation_id": 141, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4133e0f-55", "ovs_interfaceid": "e4133e0f-55c1-4b3f-847d-899dd1167eec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 794.689685] env[63531]: DEBUG oslo_vmware.api [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1117783, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.103099} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.689999] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 794.690838] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ca77c73-e9b9-4bad-abcc-dec24f12ffb9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.714238] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Reconfiguring VM instance instance-00000037 to attach disk [datastore2] 48043b5d-c0d3-4e8a-8d28-5a408d800e11/48043b5d-c0d3-4e8a-8d28-5a408d800e11.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 794.714616] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dd5dc5ea-0761-4b53-8a48-38515afcc1f4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.737507] env[63531]: DEBUG oslo_vmware.api [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for the task: (returnval){ [ 794.737507] env[63531]: value = "task-1117784" [ 794.737507] env[63531]: _type = "Task" [ 794.737507] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.746177] env[63531]: DEBUG oslo_vmware.api [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1117784, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.039417] env[63531]: DEBUG oslo_concurrency.lockutils [None req-fdd04ada-cc1b-4e03-9008-da80c3e4e411 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.066s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.042771] env[63531]: DEBUG oslo_concurrency.lockutils [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.066s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 795.046257] env[63531]: INFO nova.compute.claims [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 795.052247] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-0acd5bb8-82ed-4cf7-9ea3-96d802548d79 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Creating Snapshot of the VM instance {{(pid=63531) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 795.052996] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-04c08eff-4fbf-48b6-856c-4b0c2ba1ea92 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.064174] env[63531]: DEBUG oslo_vmware.api [None req-0acd5bb8-82ed-4cf7-9ea3-96d802548d79 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Waiting for the task: (returnval){ [ 795.064174] env[63531]: value = "task-1117785" [ 795.064174] env[63531]: _type = "Task" [ 795.064174] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.076200] env[63531]: DEBUG oslo_vmware.api [None req-0acd5bb8-82ed-4cf7-9ea3-96d802548d79 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1117785, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.084590] env[63531]: INFO nova.scheduler.client.report [None req-fdd04ada-cc1b-4e03-9008-da80c3e4e411 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Deleted allocations for instance f39c8807-1e15-41cf-899b-a1fbe0695d58 [ 795.127069] env[63531]: DEBUG oslo_concurrency.lockutils [req-118de9d7-f22e-44f3-abcd-9d2280330154 req-32807e4e-4b97-4db3-b688-fa45357659e9 service nova] Releasing lock "refresh_cache-9a3fad46-a15a-451c-bdab-a3c8cc8add07" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 795.255633] env[63531]: DEBUG oslo_vmware.api [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1117784, 'name': ReconfigVM_Task, 'duration_secs': 0.331074} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.255988] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Reconfigured VM instance instance-00000037 to attach disk [datastore2] 48043b5d-c0d3-4e8a-8d28-5a408d800e11/48043b5d-c0d3-4e8a-8d28-5a408d800e11.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 795.256745] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c03f4c02-f90f-4bd5-a1d6-e9f79bd311d4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.268471] env[63531]: DEBUG oslo_vmware.api [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for the task: (returnval){ [ 795.268471] env[63531]: value = "task-1117786" [ 795.268471] env[63531]: _type = "Task" [ 795.268471] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.280509] env[63531]: DEBUG oslo_vmware.api [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1117786, 'name': Rename_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.577140] env[63531]: DEBUG oslo_vmware.api [None req-0acd5bb8-82ed-4cf7-9ea3-96d802548d79 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1117785, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.592353] env[63531]: DEBUG oslo_concurrency.lockutils [None req-fdd04ada-cc1b-4e03-9008-da80c3e4e411 tempest-ListServersNegativeTestJSON-959341556 tempest-ListServersNegativeTestJSON-959341556-project-member] Lock "f39c8807-1e15-41cf-899b-a1fbe0695d58" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.412s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.780360] env[63531]: DEBUG oslo_vmware.api [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1117786, 'name': Rename_Task, 'duration_secs': 0.305548} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.780726] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 795.781523] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-57080f0f-99cd-4b74-b1e0-53d5c0307776 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.789246] env[63531]: DEBUG oslo_vmware.api [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for the task: (returnval){ [ 795.789246] env[63531]: value = "task-1117787" [ 795.789246] env[63531]: _type = "Task" [ 795.789246] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.807446] env[63531]: DEBUG oslo_vmware.api [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1117787, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.917429] env[63531]: DEBUG oslo_concurrency.lockutils [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Acquiring lock "56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 795.917429] env[63531]: DEBUG oslo_concurrency.lockutils [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Lock "56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.074730] env[63531]: DEBUG oslo_vmware.api [None req-0acd5bb8-82ed-4cf7-9ea3-96d802548d79 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1117785, 'name': CreateSnapshot_Task, 'duration_secs': 0.957438} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.075440] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-0acd5bb8-82ed-4cf7-9ea3-96d802548d79 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Created Snapshot of the VM instance {{(pid=63531) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 796.076890] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3264bb9c-4ff2-412f-8dd0-ac86bda62885 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.308662] env[63531]: DEBUG oslo_vmware.api [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1117787, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.421850] env[63531]: DEBUG nova.compute.manager [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] [instance: 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 796.453891] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa8b5918-4272-421c-892b-baf2c9178f95 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.463754] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9203828a-28f7-4771-9c8b-6192f9d53152 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.500694] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-885d0aac-2c19-4efb-9c4e-cc567c253c71 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.508619] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b47ee11-95e8-4c34-8322-6f3260406d8d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.522740] env[63531]: DEBUG nova.compute.provider_tree [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 796.597980] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-0acd5bb8-82ed-4cf7-9ea3-96d802548d79 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Creating linked-clone VM from snapshot {{(pid=63531) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 796.599426] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-43d085ab-4646-4aea-b3d9-f591ba694cf5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.608628] env[63531]: DEBUG oslo_vmware.api [None req-0acd5bb8-82ed-4cf7-9ea3-96d802548d79 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Waiting for the task: (returnval){ [ 796.608628] env[63531]: value = "task-1117788" [ 796.608628] env[63531]: _type = "Task" [ 796.608628] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.617828] env[63531]: DEBUG oslo_vmware.api [None req-0acd5bb8-82ed-4cf7-9ea3-96d802548d79 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1117788, 'name': CloneVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.801403] env[63531]: DEBUG oslo_vmware.api [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1117787, 'name': PowerOnVM_Task, 'duration_secs': 0.551251} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.801884] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 796.802165] env[63531]: INFO nova.compute.manager [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Took 7.67 seconds to spawn the instance on the hypervisor. [ 796.803460] env[63531]: DEBUG nova.compute.manager [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 796.803460] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8476aa2c-b143-4ebd-b8ae-dd78f7e5c84a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.957132] env[63531]: DEBUG oslo_concurrency.lockutils [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 797.026036] env[63531]: DEBUG nova.scheduler.client.report [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 797.123312] env[63531]: DEBUG oslo_vmware.api [None req-0acd5bb8-82ed-4cf7-9ea3-96d802548d79 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1117788, 'name': CloneVM_Task} progress is 94%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.330745] env[63531]: INFO nova.compute.manager [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Took 39.50 seconds to build instance. [ 797.532105] env[63531]: DEBUG oslo_concurrency.lockutils [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.489s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 797.532750] env[63531]: DEBUG nova.compute.manager [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 797.536050] env[63531]: DEBUG oslo_concurrency.lockutils [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.059s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 797.538992] env[63531]: INFO nova.compute.claims [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 797.623878] env[63531]: DEBUG oslo_vmware.api [None req-0acd5bb8-82ed-4cf7-9ea3-96d802548d79 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1117788, 'name': CloneVM_Task} progress is 94%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.834579] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c96525c0-48b0-47e0-8633-374368bf9462 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "48043b5d-c0d3-4e8a-8d28-5a408d800e11" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 101.936s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 798.045951] env[63531]: DEBUG nova.compute.utils [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 798.050406] env[63531]: DEBUG nova.compute.manager [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 798.050784] env[63531]: DEBUG nova.network.neutron [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 798.124569] env[63531]: DEBUG oslo_vmware.api [None req-0acd5bb8-82ed-4cf7-9ea3-96d802548d79 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1117788, 'name': CloneVM_Task} progress is 95%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.128566] env[63531]: DEBUG nova.policy [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '268494bb8db84ede854304a23497d130', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a0e384d6cf184a98aa3602eca6b3f410', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 798.353468] env[63531]: DEBUG nova.compute.manager [req-b4ea85f0-f73e-4ceb-96ba-5e9722a09b4b req-8c46179f-6853-486a-9886-a7c5fe40fea7 service nova] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Received event network-changed-44f13e67-ffdf-401f-bdfe-fbd57e4f4d63 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 798.353697] env[63531]: DEBUG nova.compute.manager [req-b4ea85f0-f73e-4ceb-96ba-5e9722a09b4b req-8c46179f-6853-486a-9886-a7c5fe40fea7 service nova] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Refreshing instance network info cache due to event network-changed-44f13e67-ffdf-401f-bdfe-fbd57e4f4d63. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 798.355185] env[63531]: DEBUG oslo_concurrency.lockutils [req-b4ea85f0-f73e-4ceb-96ba-5e9722a09b4b req-8c46179f-6853-486a-9886-a7c5fe40fea7 service nova] Acquiring lock "refresh_cache-48043b5d-c0d3-4e8a-8d28-5a408d800e11" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 798.355786] env[63531]: DEBUG oslo_concurrency.lockutils [req-b4ea85f0-f73e-4ceb-96ba-5e9722a09b4b req-8c46179f-6853-486a-9886-a7c5fe40fea7 service nova] Acquired lock "refresh_cache-48043b5d-c0d3-4e8a-8d28-5a408d800e11" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.355786] env[63531]: DEBUG nova.network.neutron [req-b4ea85f0-f73e-4ceb-96ba-5e9722a09b4b req-8c46179f-6853-486a-9886-a7c5fe40fea7 service nova] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Refreshing network info cache for port 44f13e67-ffdf-401f-bdfe-fbd57e4f4d63 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 798.449989] env[63531]: DEBUG nova.network.neutron [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Successfully created port: d99cda7f-476b-4869-8ba7-4a00c03da394 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 798.556397] env[63531]: DEBUG nova.compute.manager [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 798.624656] env[63531]: DEBUG oslo_vmware.api [None req-0acd5bb8-82ed-4cf7-9ea3-96d802548d79 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1117788, 'name': CloneVM_Task, 'duration_secs': 1.676942} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.628643] env[63531]: INFO nova.virt.vmwareapi.vmops [None req-0acd5bb8-82ed-4cf7-9ea3-96d802548d79 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Created linked-clone VM from snapshot [ 798.629721] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-480f6405-4854-42fb-8f07-26edba2d6e0c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.638919] env[63531]: DEBUG nova.virt.vmwareapi.images [None req-0acd5bb8-82ed-4cf7-9ea3-96d802548d79 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Uploading image 8197d52e-21f1-413d-8ef3-59307facd73d {{(pid=63531) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 798.665185] env[63531]: DEBUG oslo_vmware.rw_handles [None req-0acd5bb8-82ed-4cf7-9ea3-96d802548d79 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 798.665185] env[63531]: value = "vm-244671" [ 798.665185] env[63531]: _type = "VirtualMachine" [ 798.665185] env[63531]: }. {{(pid=63531) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 798.666479] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-c3055577-c3af-445c-b3ba-62ce337cb851 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.673601] env[63531]: DEBUG oslo_vmware.rw_handles [None req-0acd5bb8-82ed-4cf7-9ea3-96d802548d79 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Lease: (returnval){ [ 798.673601] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5247c1fd-d771-2354-4628-ac63311d0e4f" [ 798.673601] env[63531]: _type = "HttpNfcLease" [ 798.673601] env[63531]: } obtained for exporting VM: (result){ [ 798.673601] env[63531]: value = "vm-244671" [ 798.673601] env[63531]: _type = "VirtualMachine" [ 798.673601] env[63531]: }. {{(pid=63531) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 798.673886] env[63531]: DEBUG oslo_vmware.api [None req-0acd5bb8-82ed-4cf7-9ea3-96d802548d79 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Waiting for the lease: (returnval){ [ 798.673886] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5247c1fd-d771-2354-4628-ac63311d0e4f" [ 798.673886] env[63531]: _type = "HttpNfcLease" [ 798.673886] env[63531]: } to be ready. {{(pid=63531) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 798.684646] env[63531]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 798.684646] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5247c1fd-d771-2354-4628-ac63311d0e4f" [ 798.684646] env[63531]: _type = "HttpNfcLease" [ 798.684646] env[63531]: } is initializing. {{(pid=63531) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 798.959139] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d9b5588-d682-46b4-a5b5-ee0cb145ce2c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.971021] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05b6cc87-0b83-4fc9-833a-31c3aeab40ef {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.011760] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bbe2bd3-2e20-4354-97d1-64822cb1c703 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.020603] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74b2ddce-5ede-445c-ac16-19cd52ae3166 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.037590] env[63531]: DEBUG nova.compute.provider_tree [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 799.183041] env[63531]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 799.183041] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5247c1fd-d771-2354-4628-ac63311d0e4f" [ 799.183041] env[63531]: _type = "HttpNfcLease" [ 799.183041] env[63531]: } is ready. {{(pid=63531) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 799.183041] env[63531]: DEBUG oslo_vmware.rw_handles [None req-0acd5bb8-82ed-4cf7-9ea3-96d802548d79 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 799.183041] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5247c1fd-d771-2354-4628-ac63311d0e4f" [ 799.183041] env[63531]: _type = "HttpNfcLease" [ 799.183041] env[63531]: }. {{(pid=63531) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 799.183539] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78ddf5bb-ff97-43cd-bdf5-35cb85a77404 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.193319] env[63531]: DEBUG oslo_vmware.rw_handles [None req-0acd5bb8-82ed-4cf7-9ea3-96d802548d79 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ac0d79-036e-727d-07c8-606921c098df/disk-0.vmdk from lease info. {{(pid=63531) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 799.193319] env[63531]: DEBUG oslo_vmware.rw_handles [None req-0acd5bb8-82ed-4cf7-9ea3-96d802548d79 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ac0d79-036e-727d-07c8-606921c098df/disk-0.vmdk for reading. {{(pid=63531) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 799.256935] env[63531]: DEBUG nova.network.neutron [req-b4ea85f0-f73e-4ceb-96ba-5e9722a09b4b req-8c46179f-6853-486a-9886-a7c5fe40fea7 service nova] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Updated VIF entry in instance network info cache for port 44f13e67-ffdf-401f-bdfe-fbd57e4f4d63. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 799.258740] env[63531]: DEBUG nova.network.neutron [req-b4ea85f0-f73e-4ceb-96ba-5e9722a09b4b req-8c46179f-6853-486a-9886-a7c5fe40fea7 service nova] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Updating instance_info_cache with network_info: [{"id": "44f13e67-ffdf-401f-bdfe-fbd57e4f4d63", "address": "fa:16:3e:c5:55:ae", "network": {"id": "95ed2c7d-3d1a-4736-8e90-104002f5c40b", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-73549344-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.229", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9056486e3b640d0bcdb9369ed5d67ad", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04e15990-16e1-4cb2-b0f0-06c362e68c5e", "external-id": "nsx-vlan-transportzone-555", "segmentation_id": 555, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44f13e67-ff", "ovs_interfaceid": "44f13e67-ffdf-401f-bdfe-fbd57e4f4d63", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 799.293308] env[63531]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-66229da2-f211-4111-8a70-58068cf9c9ea {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.540938] env[63531]: DEBUG nova.scheduler.client.report [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 799.570604] env[63531]: DEBUG nova.compute.manager [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 799.607492] env[63531]: DEBUG nova.virt.hardware [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 799.607704] env[63531]: DEBUG nova.virt.hardware [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 799.607892] env[63531]: DEBUG nova.virt.hardware [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 799.608111] env[63531]: DEBUG nova.virt.hardware [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 799.608272] env[63531]: DEBUG nova.virt.hardware [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 799.608423] env[63531]: DEBUG nova.virt.hardware [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 799.608645] env[63531]: DEBUG nova.virt.hardware [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 799.608836] env[63531]: DEBUG nova.virt.hardware [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 799.609498] env[63531]: DEBUG nova.virt.hardware [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 799.609498] env[63531]: DEBUG nova.virt.hardware [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 799.609498] env[63531]: DEBUG nova.virt.hardware [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 799.610424] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a6ba908-ab9f-4baa-9f3b-ad5fd0471677 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.619964] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0b3e006-9e57-4725-b318-8fba38b89680 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.762463] env[63531]: DEBUG oslo_concurrency.lockutils [req-b4ea85f0-f73e-4ceb-96ba-5e9722a09b4b req-8c46179f-6853-486a-9886-a7c5fe40fea7 service nova] Releasing lock "refresh_cache-48043b5d-c0d3-4e8a-8d28-5a408d800e11" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 799.940498] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Acquiring lock "3eeaa6f9-60d9-43fe-a438-fbf2ceececae" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 799.940810] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Lock "3eeaa6f9-60d9-43fe-a438-fbf2ceececae" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.053705] env[63531]: DEBUG oslo_concurrency.lockutils [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.515s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.053705] env[63531]: DEBUG nova.compute.manager [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 800.056562] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.119s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.058887] env[63531]: INFO nova.compute.claims [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 531e1852-0c67-4b4d-a0ca-749ac438e79a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 800.072147] env[63531]: DEBUG nova.network.neutron [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Successfully updated port: d99cda7f-476b-4869-8ba7-4a00c03da394 {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 800.337415] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Acquiring lock "09feb12a-da9f-4bb7-959d-84c5c7c58e99" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 800.338338] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Lock "09feb12a-da9f-4bb7-959d-84c5c7c58e99" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.444194] env[63531]: DEBUG nova.compute.manager [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 3eeaa6f9-60d9-43fe-a438-fbf2ceececae] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 800.573482] env[63531]: DEBUG nova.compute.utils [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 800.579128] env[63531]: DEBUG nova.compute.manager [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 800.579327] env[63531]: DEBUG nova.network.neutron [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 800.582989] env[63531]: DEBUG oslo_concurrency.lockutils [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Acquiring lock "refresh_cache-caf0dee8-cb5a-402b-b533-22a590433f54" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 800.583251] env[63531]: DEBUG oslo_concurrency.lockutils [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Acquired lock "refresh_cache-caf0dee8-cb5a-402b-b533-22a590433f54" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.583885] env[63531]: DEBUG nova.network.neutron [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 800.660619] env[63531]: DEBUG nova.policy [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2f0c6adf85584641a2c6d14f56f10674', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ef3b56ce3f3a4155bd09642f5c48ea6f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 800.974563] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 801.080411] env[63531]: DEBUG nova.compute.manager [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 801.136912] env[63531]: DEBUG nova.network.neutron [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 801.213391] env[63531]: DEBUG oslo_vmware.rw_handles [None req-1171cdb2-37ed-4637-9a88-be822ac0cef5 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f30212-f5b0-861d-c7c3-797189e87aa9/disk-0.vmdk. {{(pid=63531) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 801.213391] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4ebbfd4-1e80-4341-bcbb-2cd94575c363 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.220022] env[63531]: DEBUG oslo_vmware.rw_handles [None req-1171cdb2-37ed-4637-9a88-be822ac0cef5 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f30212-f5b0-861d-c7c3-797189e87aa9/disk-0.vmdk is in state: ready. {{(pid=63531) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 801.220022] env[63531]: ERROR oslo_vmware.rw_handles [None req-1171cdb2-37ed-4637-9a88-be822ac0cef5 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f30212-f5b0-861d-c7c3-797189e87aa9/disk-0.vmdk due to incomplete transfer. [ 801.220022] env[63531]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-5e6819af-1759-4b90-98f9-06802aa315f5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.229848] env[63531]: DEBUG oslo_vmware.rw_handles [None req-1171cdb2-37ed-4637-9a88-be822ac0cef5 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f30212-f5b0-861d-c7c3-797189e87aa9/disk-0.vmdk. {{(pid=63531) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 801.229972] env[63531]: DEBUG nova.virt.vmwareapi.images [None req-1171cdb2-37ed-4637-9a88-be822ac0cef5 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Uploaded image c172132b-5ac9-4ddb-885b-cb2c5fed2c85 to the Glance image server {{(pid=63531) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 801.233407] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-1171cdb2-37ed-4637-9a88-be822ac0cef5 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Destroying the VM {{(pid=63531) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 801.233734] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-718c2ed3-8e73-4789-8fe5-dd6858b3708a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.237325] env[63531]: DEBUG nova.network.neutron [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Successfully created port: 2d27a8a5-c288-44fe-ac09-0d3b488386d6 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 801.241359] env[63531]: DEBUG oslo_vmware.api [None req-1171cdb2-37ed-4637-9a88-be822ac0cef5 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Waiting for the task: (returnval){ [ 801.241359] env[63531]: value = "task-1117790" [ 801.241359] env[63531]: _type = "Task" [ 801.241359] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.252472] env[63531]: DEBUG oslo_vmware.api [None req-1171cdb2-37ed-4637-9a88-be822ac0cef5 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117790, 'name': Destroy_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.317374] env[63531]: DEBUG nova.compute.manager [req-0a003513-6a41-4065-91e5-60be33b7de2c req-7382e8a0-e46c-4c81-96d4-c5ad4334ec58 service nova] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Received event network-vif-plugged-d99cda7f-476b-4869-8ba7-4a00c03da394 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 801.317601] env[63531]: DEBUG oslo_concurrency.lockutils [req-0a003513-6a41-4065-91e5-60be33b7de2c req-7382e8a0-e46c-4c81-96d4-c5ad4334ec58 service nova] Acquiring lock "caf0dee8-cb5a-402b-b533-22a590433f54-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 801.317912] env[63531]: DEBUG oslo_concurrency.lockutils [req-0a003513-6a41-4065-91e5-60be33b7de2c req-7382e8a0-e46c-4c81-96d4-c5ad4334ec58 service nova] Lock "caf0dee8-cb5a-402b-b533-22a590433f54-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 801.317967] env[63531]: DEBUG oslo_concurrency.lockutils [req-0a003513-6a41-4065-91e5-60be33b7de2c req-7382e8a0-e46c-4c81-96d4-c5ad4334ec58 service nova] Lock "caf0dee8-cb5a-402b-b533-22a590433f54-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 801.318187] env[63531]: DEBUG nova.compute.manager [req-0a003513-6a41-4065-91e5-60be33b7de2c req-7382e8a0-e46c-4c81-96d4-c5ad4334ec58 service nova] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] No waiting events found dispatching network-vif-plugged-d99cda7f-476b-4869-8ba7-4a00c03da394 {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 801.318359] env[63531]: WARNING nova.compute.manager [req-0a003513-6a41-4065-91e5-60be33b7de2c req-7382e8a0-e46c-4c81-96d4-c5ad4334ec58 service nova] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Received unexpected event network-vif-plugged-d99cda7f-476b-4869-8ba7-4a00c03da394 for instance with vm_state building and task_state spawning. [ 801.392219] env[63531]: DEBUG nova.network.neutron [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Updating instance_info_cache with network_info: [{"id": "d99cda7f-476b-4869-8ba7-4a00c03da394", "address": "fa:16:3e:9d:85:ab", "network": {"id": "067f0236-aa59-4a63-a33c-455dc28e0c1a", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-54491452-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a0e384d6cf184a98aa3602eca6b3f410", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7bdf594e-da7a-4254-b413-87aef4614588", "external-id": "nsx-vlan-transportzone-422", "segmentation_id": 422, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd99cda7f-47", "ovs_interfaceid": "d99cda7f-476b-4869-8ba7-4a00c03da394", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.533162] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1f2eb95-d9cf-43f2-b480-d3158c52b6e2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.541553] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e8a5b5b-b08b-490d-8d85-ccc748eb23a9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.574371] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7224fe77-bc59-4226-8571-acb1585698e7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.583800] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53f8d5c3-17f9-454a-8177-48f38a4d9379 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.605591] env[63531]: DEBUG nova.compute.provider_tree [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 801.752718] env[63531]: DEBUG oslo_vmware.api [None req-1171cdb2-37ed-4637-9a88-be822ac0cef5 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117790, 'name': Destroy_Task} progress is 33%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.897436] env[63531]: DEBUG oslo_concurrency.lockutils [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Releasing lock "refresh_cache-caf0dee8-cb5a-402b-b533-22a590433f54" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.897845] env[63531]: DEBUG nova.compute.manager [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Instance network_info: |[{"id": "d99cda7f-476b-4869-8ba7-4a00c03da394", "address": "fa:16:3e:9d:85:ab", "network": {"id": "067f0236-aa59-4a63-a33c-455dc28e0c1a", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-54491452-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a0e384d6cf184a98aa3602eca6b3f410", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7bdf594e-da7a-4254-b413-87aef4614588", "external-id": "nsx-vlan-transportzone-422", "segmentation_id": 422, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd99cda7f-47", "ovs_interfaceid": "d99cda7f-476b-4869-8ba7-4a00c03da394", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 801.898508] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9d:85:ab', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7bdf594e-da7a-4254-b413-87aef4614588', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd99cda7f-476b-4869-8ba7-4a00c03da394', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 801.909096] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Creating folder: Project (a0e384d6cf184a98aa3602eca6b3f410). Parent ref: group-v244585. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 801.909096] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3ff1c6ed-030e-4ed0-97b6-5d2d7f888ede {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.920649] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Created folder: Project (a0e384d6cf184a98aa3602eca6b3f410) in parent group-v244585. [ 801.920729] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Creating folder: Instances. Parent ref: group-v244672. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 801.921276] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b4015877-cefa-4679-bf5c-4d40ac3fbc19 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.932068] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Created folder: Instances in parent group-v244672. [ 801.932363] env[63531]: DEBUG oslo.service.loopingcall [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 801.932626] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 801.932960] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-28b43d15-56f7-48c4-b0ae-417d6a43b7b8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.955602] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 801.955602] env[63531]: value = "task-1117793" [ 801.955602] env[63531]: _type = "Task" [ 801.955602] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.965902] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117793, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.094610] env[63531]: DEBUG nova.compute.manager [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 802.112882] env[63531]: DEBUG nova.scheduler.client.report [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 802.128778] env[63531]: DEBUG nova.virt.hardware [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 802.129040] env[63531]: DEBUG nova.virt.hardware [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 802.129207] env[63531]: DEBUG nova.virt.hardware [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 802.129398] env[63531]: DEBUG nova.virt.hardware [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 802.129558] env[63531]: DEBUG nova.virt.hardware [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 802.129702] env[63531]: DEBUG nova.virt.hardware [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 802.129927] env[63531]: DEBUG nova.virt.hardware [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 802.130543] env[63531]: DEBUG nova.virt.hardware [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 802.130786] env[63531]: DEBUG nova.virt.hardware [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 802.130969] env[63531]: DEBUG nova.virt.hardware [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 802.131504] env[63531]: DEBUG nova.virt.hardware [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 802.132057] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b18386bd-b2c6-40f8-9c02-902f544d3399 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.142650] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2439da8-ddcc-4198-b2de-53cfef03877f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.252367] env[63531]: DEBUG oslo_vmware.api [None req-1171cdb2-37ed-4637-9a88-be822ac0cef5 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117790, 'name': Destroy_Task, 'duration_secs': 0.605287} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.252579] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-1171cdb2-37ed-4637-9a88-be822ac0cef5 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Destroyed the VM [ 802.252916] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-1171cdb2-37ed-4637-9a88-be822ac0cef5 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Deleting Snapshot of the VM instance {{(pid=63531) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 802.253286] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-8a7f5826-1ff1-4e1e-aaf3-f4035b94cf48 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.261017] env[63531]: DEBUG oslo_vmware.api [None req-1171cdb2-37ed-4637-9a88-be822ac0cef5 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Waiting for the task: (returnval){ [ 802.261017] env[63531]: value = "task-1117794" [ 802.261017] env[63531]: _type = "Task" [ 802.261017] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.268795] env[63531]: DEBUG oslo_vmware.api [None req-1171cdb2-37ed-4637-9a88-be822ac0cef5 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117794, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.470673] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117793, 'name': CreateVM_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.622945] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.567s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.623671] env[63531]: DEBUG nova.compute.manager [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 531e1852-0c67-4b4d-a0ca-749ac438e79a] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 802.626879] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7385c1eb-e474-478e-a87d-01f2ddd50615 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.207s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 802.627180] env[63531]: DEBUG nova.objects.instance [None req-7385c1eb-e474-478e-a87d-01f2ddd50615 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Lazy-loading 'resources' on Instance uuid e1693594-d8bb-4002-b71c-7013efb81d16 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 802.778152] env[63531]: DEBUG oslo_vmware.api [None req-1171cdb2-37ed-4637-9a88-be822ac0cef5 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117794, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.967098] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117793, 'name': CreateVM_Task, 'duration_secs': 0.537479} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.967301] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 802.968037] env[63531]: DEBUG oslo_concurrency.lockutils [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.968239] env[63531]: DEBUG oslo_concurrency.lockutils [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.968563] env[63531]: DEBUG oslo_concurrency.lockutils [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 802.968821] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-37c963de-7798-4883-b933-d80439739ddf {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.974124] env[63531]: DEBUG oslo_vmware.api [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Waiting for the task: (returnval){ [ 802.974124] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5243439f-a88c-da7f-0141-325973ac1bc4" [ 802.974124] env[63531]: _type = "Task" [ 802.974124] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.982459] env[63531]: DEBUG oslo_vmware.api [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5243439f-a88c-da7f-0141-325973ac1bc4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.043024] env[63531]: DEBUG nova.network.neutron [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Successfully updated port: 2d27a8a5-c288-44fe-ac09-0d3b488386d6 {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 803.130448] env[63531]: DEBUG nova.compute.utils [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 803.138134] env[63531]: DEBUG nova.compute.manager [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 531e1852-0c67-4b4d-a0ca-749ac438e79a] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 803.138134] env[63531]: DEBUG nova.network.neutron [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 531e1852-0c67-4b4d-a0ca-749ac438e79a] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 803.195743] env[63531]: DEBUG nova.policy [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7965dce40de84f8083d57a8d51d4a063', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f727ddb992dc4c06954f25f6a1613b4a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 803.285450] env[63531]: DEBUG oslo_vmware.api [None req-1171cdb2-37ed-4637-9a88-be822ac0cef5 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117794, 'name': RemoveSnapshot_Task, 'duration_secs': 0.762105} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.285450] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-1171cdb2-37ed-4637-9a88-be822ac0cef5 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Deleted Snapshot of the VM instance {{(pid=63531) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 803.285450] env[63531]: INFO nova.compute.manager [None req-1171cdb2-37ed-4637-9a88-be822ac0cef5 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Took 17.21 seconds to snapshot the instance on the hypervisor. [ 803.366632] env[63531]: DEBUG nova.compute.manager [req-fb78a317-d912-429f-920d-129e880ace0e req-9a146bc3-a1f8-414c-8822-ac3c0a893747 service nova] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Received event network-changed-d99cda7f-476b-4869-8ba7-4a00c03da394 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 803.366917] env[63531]: DEBUG nova.compute.manager [req-fb78a317-d912-429f-920d-129e880ace0e req-9a146bc3-a1f8-414c-8822-ac3c0a893747 service nova] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Refreshing instance network info cache due to event network-changed-d99cda7f-476b-4869-8ba7-4a00c03da394. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 803.368966] env[63531]: DEBUG oslo_concurrency.lockutils [req-fb78a317-d912-429f-920d-129e880ace0e req-9a146bc3-a1f8-414c-8822-ac3c0a893747 service nova] Acquiring lock "refresh_cache-caf0dee8-cb5a-402b-b533-22a590433f54" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 803.368966] env[63531]: DEBUG oslo_concurrency.lockutils [req-fb78a317-d912-429f-920d-129e880ace0e req-9a146bc3-a1f8-414c-8822-ac3c0a893747 service nova] Acquired lock "refresh_cache-caf0dee8-cb5a-402b-b533-22a590433f54" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 803.368966] env[63531]: DEBUG nova.network.neutron [req-fb78a317-d912-429f-920d-129e880ace0e req-9a146bc3-a1f8-414c-8822-ac3c0a893747 service nova] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Refreshing network info cache for port d99cda7f-476b-4869-8ba7-4a00c03da394 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 803.494181] env[63531]: DEBUG oslo_vmware.api [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5243439f-a88c-da7f-0141-325973ac1bc4, 'name': SearchDatastore_Task, 'duration_secs': 0.017486} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.498737] env[63531]: DEBUG oslo_concurrency.lockutils [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 803.499096] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 803.499415] env[63531]: DEBUG oslo_concurrency.lockutils [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 803.499941] env[63531]: DEBUG oslo_concurrency.lockutils [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 803.499941] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 803.500954] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1f97d493-b5b2-4859-8847-b57e35cdef70 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.512306] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 803.513381] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 803.513381] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39d9d01e-9326-445e-a8b7-08c20253cbee {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.518787] env[63531]: DEBUG nova.network.neutron [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 531e1852-0c67-4b4d-a0ca-749ac438e79a] Successfully created port: 99ca461b-afde-4ea5-8cbc-e596f9f8e528 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 803.524338] env[63531]: DEBUG oslo_vmware.api [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Waiting for the task: (returnval){ [ 803.524338] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52a5a13f-515e-ac1f-9972-5dd4aa9ffb15" [ 803.524338] env[63531]: _type = "Task" [ 803.524338] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.536545] env[63531]: DEBUG oslo_vmware.api [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52a5a13f-515e-ac1f-9972-5dd4aa9ffb15, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.547685] env[63531]: DEBUG oslo_concurrency.lockutils [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Acquiring lock "refresh_cache-2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 803.548039] env[63531]: DEBUG oslo_concurrency.lockutils [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Acquired lock "refresh_cache-2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 803.548114] env[63531]: DEBUG nova.network.neutron [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 803.639461] env[63531]: DEBUG nova.compute.manager [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 531e1852-0c67-4b4d-a0ca-749ac438e79a] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 803.709791] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ec62ceb-55d2-40f5-9060-82f3e20550e3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.718499] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9dbe0b5-634e-41d5-9ae2-121b62b20b9b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.748863] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ee94f8d-db10-4b4a-be8d-1d978eafe983 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.758801] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-817b64a6-12d5-4c93-9bf2-6bb5cefde73e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.771082] env[63531]: DEBUG nova.compute.provider_tree [None req-7385c1eb-e474-478e-a87d-01f2ddd50615 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 804.010189] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "2b0e4285-310a-439c-bc50-000a7c5ef7f9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 804.010466] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "2b0e4285-310a-439c-bc50-000a7c5ef7f9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 804.037766] env[63531]: DEBUG oslo_vmware.api [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52a5a13f-515e-ac1f-9972-5dd4aa9ffb15, 'name': SearchDatastore_Task, 'duration_secs': 0.016158} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.038742] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e75eacd6-5f91-49fc-ada6-69a8bd8e39d3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.045457] env[63531]: DEBUG oslo_vmware.api [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Waiting for the task: (returnval){ [ 804.045457] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52d9594e-963b-2f18-eeb5-44f8885d03e6" [ 804.045457] env[63531]: _type = "Task" [ 804.045457] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.061142] env[63531]: DEBUG oslo_vmware.api [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52d9594e-963b-2f18-eeb5-44f8885d03e6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.094952] env[63531]: DEBUG nova.network.neutron [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 804.263851] env[63531]: DEBUG nova.network.neutron [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Updating instance_info_cache with network_info: [{"id": "2d27a8a5-c288-44fe-ac09-0d3b488386d6", "address": "fa:16:3e:91:1c:62", "network": {"id": "5b8b7134-f123-40b8-9134-5cd072e8ad01", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1232565840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef3b56ce3f3a4155bd09642f5c48ea6f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2d27a8a5-c2", "ovs_interfaceid": "2d27a8a5-c288-44fe-ac09-0d3b488386d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.276289] env[63531]: DEBUG nova.scheduler.client.report [None req-7385c1eb-e474-478e-a87d-01f2ddd50615 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 804.321899] env[63531]: DEBUG nova.network.neutron [req-fb78a317-d912-429f-920d-129e880ace0e req-9a146bc3-a1f8-414c-8822-ac3c0a893747 service nova] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Updated VIF entry in instance network info cache for port d99cda7f-476b-4869-8ba7-4a00c03da394. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 804.322412] env[63531]: DEBUG nova.network.neutron [req-fb78a317-d912-429f-920d-129e880ace0e req-9a146bc3-a1f8-414c-8822-ac3c0a893747 service nova] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Updating instance_info_cache with network_info: [{"id": "d99cda7f-476b-4869-8ba7-4a00c03da394", "address": "fa:16:3e:9d:85:ab", "network": {"id": "067f0236-aa59-4a63-a33c-455dc28e0c1a", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-54491452-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a0e384d6cf184a98aa3602eca6b3f410", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7bdf594e-da7a-4254-b413-87aef4614588", "external-id": "nsx-vlan-transportzone-422", "segmentation_id": 422, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd99cda7f-47", "ovs_interfaceid": "d99cda7f-476b-4869-8ba7-4a00c03da394", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.557151] env[63531]: DEBUG oslo_vmware.api [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52d9594e-963b-2f18-eeb5-44f8885d03e6, 'name': SearchDatastore_Task, 'duration_secs': 0.01722} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.557615] env[63531]: DEBUG oslo_concurrency.lockutils [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.557967] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] caf0dee8-cb5a-402b-b533-22a590433f54/caf0dee8-cb5a-402b-b533-22a590433f54.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 804.558197] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-544e8655-5d5c-412f-8b06-55c4baca1a18 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.568272] env[63531]: DEBUG oslo_vmware.api [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Waiting for the task: (returnval){ [ 804.568272] env[63531]: value = "task-1117795" [ 804.568272] env[63531]: _type = "Task" [ 804.568272] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.577795] env[63531]: DEBUG oslo_vmware.api [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Task: {'id': task-1117795, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.653988] env[63531]: DEBUG nova.compute.manager [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 531e1852-0c67-4b4d-a0ca-749ac438e79a] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 804.768578] env[63531]: DEBUG oslo_concurrency.lockutils [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Releasing lock "refresh_cache-2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.768804] env[63531]: DEBUG nova.compute.manager [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Instance network_info: |[{"id": "2d27a8a5-c288-44fe-ac09-0d3b488386d6", "address": "fa:16:3e:91:1c:62", "network": {"id": "5b8b7134-f123-40b8-9134-5cd072e8ad01", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1232565840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef3b56ce3f3a4155bd09642f5c48ea6f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2d27a8a5-c2", "ovs_interfaceid": "2d27a8a5-c288-44fe-ac09-0d3b488386d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 804.769429] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:91:1c:62', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd891a662-7da0-4be5-8d0d-01b4ff70552a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2d27a8a5-c288-44fe-ac09-0d3b488386d6', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 804.777799] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Creating folder: Project (ef3b56ce3f3a4155bd09642f5c48ea6f). Parent ref: group-v244585. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 804.778144] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-df940c2e-6ec2-4be5-ac37-d73e5d27c1a4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.784460] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7385c1eb-e474-478e-a87d-01f2ddd50615 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.158s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.787700] env[63531]: DEBUG oslo_concurrency.lockutils [None req-91e9db6d-db1f-484a-9006-e9fab2e8db3e tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.876s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 804.788052] env[63531]: DEBUG nova.objects.instance [None req-91e9db6d-db1f-484a-9006-e9fab2e8db3e tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Lazy-loading 'resources' on Instance uuid dccb2335-4220-4570-861b-16ff2d8f5a85 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 804.790617] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Created folder: Project (ef3b56ce3f3a4155bd09642f5c48ea6f) in parent group-v244585. [ 804.790816] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Creating folder: Instances. Parent ref: group-v244675. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 804.791308] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8fee7637-9e9e-4109-8cfe-a3d629791d6e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.803867] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Created folder: Instances in parent group-v244675. [ 804.804340] env[63531]: DEBUG oslo.service.loopingcall [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 804.804588] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 804.804919] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1bebb4ff-62b7-48af-9a60-0eff9deb8cdd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.823971] env[63531]: INFO nova.scheduler.client.report [None req-7385c1eb-e474-478e-a87d-01f2ddd50615 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Deleted allocations for instance e1693594-d8bb-4002-b71c-7013efb81d16 [ 804.825359] env[63531]: DEBUG oslo_concurrency.lockutils [req-fb78a317-d912-429f-920d-129e880ace0e req-9a146bc3-a1f8-414c-8822-ac3c0a893747 service nova] Releasing lock "refresh_cache-caf0dee8-cb5a-402b-b533-22a590433f54" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.827500] env[63531]: DEBUG nova.compute.manager [req-fb78a317-d912-429f-920d-129e880ace0e req-9a146bc3-a1f8-414c-8822-ac3c0a893747 service nova] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Received event network-vif-plugged-2d27a8a5-c288-44fe-ac09-0d3b488386d6 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 804.827500] env[63531]: DEBUG oslo_concurrency.lockutils [req-fb78a317-d912-429f-920d-129e880ace0e req-9a146bc3-a1f8-414c-8822-ac3c0a893747 service nova] Acquiring lock "2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 804.828113] env[63531]: DEBUG oslo_concurrency.lockutils [req-fb78a317-d912-429f-920d-129e880ace0e req-9a146bc3-a1f8-414c-8822-ac3c0a893747 service nova] Lock "2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 804.828113] env[63531]: DEBUG oslo_concurrency.lockutils [req-fb78a317-d912-429f-920d-129e880ace0e req-9a146bc3-a1f8-414c-8822-ac3c0a893747 service nova] Lock "2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.828113] env[63531]: DEBUG nova.compute.manager [req-fb78a317-d912-429f-920d-129e880ace0e req-9a146bc3-a1f8-414c-8822-ac3c0a893747 service nova] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] No waiting events found dispatching network-vif-plugged-2d27a8a5-c288-44fe-ac09-0d3b488386d6 {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 804.828521] env[63531]: WARNING nova.compute.manager [req-fb78a317-d912-429f-920d-129e880ace0e req-9a146bc3-a1f8-414c-8822-ac3c0a893747 service nova] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Received unexpected event network-vif-plugged-2d27a8a5-c288-44fe-ac09-0d3b488386d6 for instance with vm_state building and task_state spawning. [ 804.828521] env[63531]: DEBUG nova.compute.manager [req-fb78a317-d912-429f-920d-129e880ace0e req-9a146bc3-a1f8-414c-8822-ac3c0a893747 service nova] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Received event network-changed-2d27a8a5-c288-44fe-ac09-0d3b488386d6 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 804.828693] env[63531]: DEBUG nova.compute.manager [req-fb78a317-d912-429f-920d-129e880ace0e req-9a146bc3-a1f8-414c-8822-ac3c0a893747 service nova] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Refreshing instance network info cache due to event network-changed-2d27a8a5-c288-44fe-ac09-0d3b488386d6. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 804.828807] env[63531]: DEBUG oslo_concurrency.lockutils [req-fb78a317-d912-429f-920d-129e880ace0e req-9a146bc3-a1f8-414c-8822-ac3c0a893747 service nova] Acquiring lock "refresh_cache-2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.828952] env[63531]: DEBUG oslo_concurrency.lockutils [req-fb78a317-d912-429f-920d-129e880ace0e req-9a146bc3-a1f8-414c-8822-ac3c0a893747 service nova] Acquired lock "refresh_cache-2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.829133] env[63531]: DEBUG nova.network.neutron [req-fb78a317-d912-429f-920d-129e880ace0e req-9a146bc3-a1f8-414c-8822-ac3c0a893747 service nova] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Refreshing network info cache for port 2d27a8a5-c288-44fe-ac09-0d3b488386d6 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 804.834442] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 804.834442] env[63531]: value = "task-1117798" [ 804.834442] env[63531]: _type = "Task" [ 804.834442] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.845149] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117798, 'name': CreateVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.883650] env[63531]: DEBUG nova.virt.hardware [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 804.885347] env[63531]: DEBUG nova.virt.hardware [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 804.885347] env[63531]: DEBUG nova.virt.hardware [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 804.885347] env[63531]: DEBUG nova.virt.hardware [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 804.885347] env[63531]: DEBUG nova.virt.hardware [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 804.885347] env[63531]: DEBUG nova.virt.hardware [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 804.885347] env[63531]: DEBUG nova.virt.hardware [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 804.885347] env[63531]: DEBUG nova.virt.hardware [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 804.885708] env[63531]: DEBUG nova.virt.hardware [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 804.885708] env[63531]: DEBUG nova.virt.hardware [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 804.885859] env[63531]: DEBUG nova.virt.hardware [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 804.887712] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7b46bd7-6a28-465e-9c56-7815d804b738 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.898263] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d046c5b1-58a8-4056-997d-21655a7440ff {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.080186] env[63531]: DEBUG oslo_vmware.api [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Task: {'id': task-1117795, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.184258] env[63531]: DEBUG nova.network.neutron [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 531e1852-0c67-4b4d-a0ca-749ac438e79a] Successfully updated port: 99ca461b-afde-4ea5-8cbc-e596f9f8e528 {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 805.342768] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7385c1eb-e474-478e-a87d-01f2ddd50615 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Lock "e1693594-d8bb-4002-b71c-7013efb81d16" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.455s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.354605] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117798, 'name': CreateVM_Task} progress is 25%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.435643] env[63531]: DEBUG nova.compute.manager [req-3427a4b0-7fdc-4268-986b-4266ed48f748 req-4eb32c89-2361-45dc-84c0-8751fe9b29f0 service nova] [instance: 531e1852-0c67-4b4d-a0ca-749ac438e79a] Received event network-vif-plugged-99ca461b-afde-4ea5-8cbc-e596f9f8e528 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 805.435862] env[63531]: DEBUG oslo_concurrency.lockutils [req-3427a4b0-7fdc-4268-986b-4266ed48f748 req-4eb32c89-2361-45dc-84c0-8751fe9b29f0 service nova] Acquiring lock "531e1852-0c67-4b4d-a0ca-749ac438e79a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 805.436326] env[63531]: DEBUG oslo_concurrency.lockutils [req-3427a4b0-7fdc-4268-986b-4266ed48f748 req-4eb32c89-2361-45dc-84c0-8751fe9b29f0 service nova] Lock "531e1852-0c67-4b4d-a0ca-749ac438e79a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 805.436543] env[63531]: DEBUG oslo_concurrency.lockutils [req-3427a4b0-7fdc-4268-986b-4266ed48f748 req-4eb32c89-2361-45dc-84c0-8751fe9b29f0 service nova] Lock "531e1852-0c67-4b4d-a0ca-749ac438e79a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.438026] env[63531]: DEBUG nova.compute.manager [req-3427a4b0-7fdc-4268-986b-4266ed48f748 req-4eb32c89-2361-45dc-84c0-8751fe9b29f0 service nova] [instance: 531e1852-0c67-4b4d-a0ca-749ac438e79a] No waiting events found dispatching network-vif-plugged-99ca461b-afde-4ea5-8cbc-e596f9f8e528 {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 805.438265] env[63531]: WARNING nova.compute.manager [req-3427a4b0-7fdc-4268-986b-4266ed48f748 req-4eb32c89-2361-45dc-84c0-8751fe9b29f0 service nova] [instance: 531e1852-0c67-4b4d-a0ca-749ac438e79a] Received unexpected event network-vif-plugged-99ca461b-afde-4ea5-8cbc-e596f9f8e528 for instance with vm_state building and task_state spawning. [ 805.438478] env[63531]: DEBUG nova.compute.manager [req-3427a4b0-7fdc-4268-986b-4266ed48f748 req-4eb32c89-2361-45dc-84c0-8751fe9b29f0 service nova] [instance: 531e1852-0c67-4b4d-a0ca-749ac438e79a] Received event network-changed-99ca461b-afde-4ea5-8cbc-e596f9f8e528 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 805.438808] env[63531]: DEBUG nova.compute.manager [req-3427a4b0-7fdc-4268-986b-4266ed48f748 req-4eb32c89-2361-45dc-84c0-8751fe9b29f0 service nova] [instance: 531e1852-0c67-4b4d-a0ca-749ac438e79a] Refreshing instance network info cache due to event network-changed-99ca461b-afde-4ea5-8cbc-e596f9f8e528. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 805.439392] env[63531]: DEBUG oslo_concurrency.lockutils [req-3427a4b0-7fdc-4268-986b-4266ed48f748 req-4eb32c89-2361-45dc-84c0-8751fe9b29f0 service nova] Acquiring lock "refresh_cache-531e1852-0c67-4b4d-a0ca-749ac438e79a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.439514] env[63531]: DEBUG oslo_concurrency.lockutils [req-3427a4b0-7fdc-4268-986b-4266ed48f748 req-4eb32c89-2361-45dc-84c0-8751fe9b29f0 service nova] Acquired lock "refresh_cache-531e1852-0c67-4b4d-a0ca-749ac438e79a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.440292] env[63531]: DEBUG nova.network.neutron [req-3427a4b0-7fdc-4268-986b-4266ed48f748 req-4eb32c89-2361-45dc-84c0-8751fe9b29f0 service nova] [instance: 531e1852-0c67-4b4d-a0ca-749ac438e79a] Refreshing network info cache for port 99ca461b-afde-4ea5-8cbc-e596f9f8e528 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 805.583852] env[63531]: DEBUG oslo_vmware.api [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Task: {'id': task-1117795, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.648478} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.587437] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] caf0dee8-cb5a-402b-b533-22a590433f54/caf0dee8-cb5a-402b-b533-22a590433f54.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 805.587724] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 805.588705] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-74f94aa0-a05c-47f4-8f5c-c5611647daa6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.597921] env[63531]: DEBUG oslo_vmware.api [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Waiting for the task: (returnval){ [ 805.597921] env[63531]: value = "task-1117799" [ 805.597921] env[63531]: _type = "Task" [ 805.597921] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.609632] env[63531]: DEBUG oslo_vmware.api [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Task: {'id': task-1117799, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.621279] env[63531]: DEBUG nova.network.neutron [req-fb78a317-d912-429f-920d-129e880ace0e req-9a146bc3-a1f8-414c-8822-ac3c0a893747 service nova] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Updated VIF entry in instance network info cache for port 2d27a8a5-c288-44fe-ac09-0d3b488386d6. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 805.621279] env[63531]: DEBUG nova.network.neutron [req-fb78a317-d912-429f-920d-129e880ace0e req-9a146bc3-a1f8-414c-8822-ac3c0a893747 service nova] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Updating instance_info_cache with network_info: [{"id": "2d27a8a5-c288-44fe-ac09-0d3b488386d6", "address": "fa:16:3e:91:1c:62", "network": {"id": "5b8b7134-f123-40b8-9134-5cd072e8ad01", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1232565840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef3b56ce3f3a4155bd09642f5c48ea6f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2d27a8a5-c2", "ovs_interfaceid": "2d27a8a5-c288-44fe-ac09-0d3b488386d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 805.688567] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Acquiring lock "refresh_cache-531e1852-0c67-4b4d-a0ca-749ac438e79a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.698985] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e16c5e9a-8f85-494a-8090-d44acff4f713 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.710449] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cf5cf3a-eb0e-4cd3-a55a-66e5939ff554 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.742294] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d4ac0fa-5741-48c0-bc00-2e74541ad886 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.750012] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c024aaec-d1c2-40c1-9a5a-e51c5d287c62 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.765426] env[63531]: DEBUG nova.compute.provider_tree [None req-91e9db6d-db1f-484a-9006-e9fab2e8db3e tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 805.852818] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117798, 'name': CreateVM_Task, 'duration_secs': 0.645283} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.853058] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 805.853817] env[63531]: DEBUG oslo_concurrency.lockutils [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.854070] env[63531]: DEBUG oslo_concurrency.lockutils [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.854476] env[63531]: DEBUG oslo_concurrency.lockutils [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 805.854790] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1ce88d26-0212-4be1-9ef4-fd1ec0086cea {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.860220] env[63531]: DEBUG oslo_vmware.api [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Waiting for the task: (returnval){ [ 805.860220] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5215219d-227f-5c68-c3a3-1f7455cebbb0" [ 805.860220] env[63531]: _type = "Task" [ 805.860220] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.868864] env[63531]: DEBUG oslo_vmware.api [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5215219d-227f-5c68-c3a3-1f7455cebbb0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.926210] env[63531]: DEBUG nova.compute.manager [None req-ad241f4e-db19-4dc3-893a-2cb7737a96b1 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 805.926916] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-740e79df-1a9a-460b-8307-5ceaa856fe80 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.001047] env[63531]: DEBUG nova.network.neutron [req-3427a4b0-7fdc-4268-986b-4266ed48f748 req-4eb32c89-2361-45dc-84c0-8751fe9b29f0 service nova] [instance: 531e1852-0c67-4b4d-a0ca-749ac438e79a] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 806.111022] env[63531]: DEBUG oslo_vmware.api [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Task: {'id': task-1117799, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.143676} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.111022] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 806.111022] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67d809f2-1319-4c09-afac-4bbb32925448 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.124128] env[63531]: DEBUG oslo_concurrency.lockutils [req-fb78a317-d912-429f-920d-129e880ace0e req-9a146bc3-a1f8-414c-8822-ac3c0a893747 service nova] Releasing lock "refresh_cache-2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.133556] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Reconfiguring VM instance instance-00000038 to attach disk [datastore1] caf0dee8-cb5a-402b-b533-22a590433f54/caf0dee8-cb5a-402b-b533-22a590433f54.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 806.134768] env[63531]: DEBUG nova.network.neutron [req-3427a4b0-7fdc-4268-986b-4266ed48f748 req-4eb32c89-2361-45dc-84c0-8751fe9b29f0 service nova] [instance: 531e1852-0c67-4b4d-a0ca-749ac438e79a] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.136031] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ce5b5922-73b2-4b2a-8f1b-111d034bf3e5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.157514] env[63531]: DEBUG oslo_vmware.api [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Waiting for the task: (returnval){ [ 806.157514] env[63531]: value = "task-1117800" [ 806.157514] env[63531]: _type = "Task" [ 806.157514] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.165881] env[63531]: DEBUG oslo_vmware.api [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Task: {'id': task-1117800, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.269037] env[63531]: DEBUG nova.scheduler.client.report [None req-91e9db6d-db1f-484a-9006-e9fab2e8db3e tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 806.372526] env[63531]: DEBUG oslo_vmware.api [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5215219d-227f-5c68-c3a3-1f7455cebbb0, 'name': SearchDatastore_Task, 'duration_secs': 0.016587} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.372882] env[63531]: DEBUG oslo_concurrency.lockutils [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.373102] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 806.373391] env[63531]: DEBUG oslo_concurrency.lockutils [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.373547] env[63531]: DEBUG oslo_concurrency.lockutils [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.373736] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 806.374013] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ffa90913-4980-4913-a16b-58d8ac9f7ca6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.389896] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 806.390117] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 806.390862] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b8ddbd6d-2e54-4a17-9d0f-0e31544bf1ac {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.399025] env[63531]: DEBUG oslo_vmware.api [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Waiting for the task: (returnval){ [ 806.399025] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52618824-fa98-da2e-c36d-08c6d7ddbba8" [ 806.399025] env[63531]: _type = "Task" [ 806.399025] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.405830] env[63531]: DEBUG oslo_vmware.api [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52618824-fa98-da2e-c36d-08c6d7ddbba8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.438827] env[63531]: INFO nova.compute.manager [None req-ad241f4e-db19-4dc3-893a-2cb7737a96b1 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] instance snapshotting [ 806.443590] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5f50323-4b32-4dff-ae28-4840fc994405 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.464136] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-165e1d60-039a-4b39-8f47-096c7d1468dd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.651870] env[63531]: DEBUG oslo_concurrency.lockutils [req-3427a4b0-7fdc-4268-986b-4266ed48f748 req-4eb32c89-2361-45dc-84c0-8751fe9b29f0 service nova] Releasing lock "refresh_cache-531e1852-0c67-4b4d-a0ca-749ac438e79a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.652310] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Acquired lock "refresh_cache-531e1852-0c67-4b4d-a0ca-749ac438e79a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.652467] env[63531]: DEBUG nova.network.neutron [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 531e1852-0c67-4b4d-a0ca-749ac438e79a] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 806.673549] env[63531]: DEBUG oslo_vmware.api [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Task: {'id': task-1117800, 'name': ReconfigVM_Task, 'duration_secs': 0.511479} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.674606] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Reconfigured VM instance instance-00000038 to attach disk [datastore1] caf0dee8-cb5a-402b-b533-22a590433f54/caf0dee8-cb5a-402b-b533-22a590433f54.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 806.675352] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2872d0a7-a0bc-4c8a-b229-584246ce2768 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.682394] env[63531]: DEBUG oslo_vmware.api [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Waiting for the task: (returnval){ [ 806.682394] env[63531]: value = "task-1117801" [ 806.682394] env[63531]: _type = "Task" [ 806.682394] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.690817] env[63531]: DEBUG oslo_vmware.api [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Task: {'id': task-1117801, 'name': Rename_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.778800] env[63531]: DEBUG oslo_concurrency.lockutils [None req-91e9db6d-db1f-484a-9006-e9fab2e8db3e tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.991s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.781216] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.445s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.782758] env[63531]: INFO nova.compute.claims [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] [instance: eb6c84d7-e0e6-4999-9018-889b99dc67ad] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 806.814306] env[63531]: INFO nova.scheduler.client.report [None req-91e9db6d-db1f-484a-9006-e9fab2e8db3e tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Deleted allocations for instance dccb2335-4220-4570-861b-16ff2d8f5a85 [ 806.911775] env[63531]: DEBUG oslo_vmware.api [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52618824-fa98-da2e-c36d-08c6d7ddbba8, 'name': SearchDatastore_Task, 'duration_secs': 0.030926} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.913120] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-12319de2-03e8-436a-aa1a-94efc1665715 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.919840] env[63531]: DEBUG oslo_vmware.api [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Waiting for the task: (returnval){ [ 806.919840] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]523f531e-acf7-719b-2e2e-db2f74cfb51b" [ 806.919840] env[63531]: _type = "Task" [ 806.919840] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.928493] env[63531]: DEBUG oslo_vmware.api [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]523f531e-acf7-719b-2e2e-db2f74cfb51b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.974655] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-ad241f4e-db19-4dc3-893a-2cb7737a96b1 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Creating Snapshot of the VM instance {{(pid=63531) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 806.975035] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-378d9177-f735-48b1-ab13-17331d5d061d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.982179] env[63531]: DEBUG oslo_vmware.api [None req-ad241f4e-db19-4dc3-893a-2cb7737a96b1 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Waiting for the task: (returnval){ [ 806.982179] env[63531]: value = "task-1117802" [ 806.982179] env[63531]: _type = "Task" [ 806.982179] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.990802] env[63531]: DEBUG oslo_vmware.api [None req-ad241f4e-db19-4dc3-893a-2cb7737a96b1 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117802, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.200643] env[63531]: DEBUG oslo_vmware.api [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Task: {'id': task-1117801, 'name': Rename_Task, 'duration_secs': 0.14525} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.201485] env[63531]: DEBUG nova.network.neutron [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 531e1852-0c67-4b4d-a0ca-749ac438e79a] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 807.203393] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 807.203761] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bb66801d-401a-419c-a559-f16d74dfcf3f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.210971] env[63531]: DEBUG oslo_vmware.api [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Waiting for the task: (returnval){ [ 807.210971] env[63531]: value = "task-1117803" [ 807.210971] env[63531]: _type = "Task" [ 807.210971] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.219964] env[63531]: DEBUG oslo_vmware.api [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Task: {'id': task-1117803, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.329043] env[63531]: DEBUG oslo_concurrency.lockutils [None req-91e9db6d-db1f-484a-9006-e9fab2e8db3e tempest-ServerMetadataTestJSON-1808803505 tempest-ServerMetadataTestJSON-1808803505-project-member] Lock "dccb2335-4220-4570-861b-16ff2d8f5a85" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.384s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.397591] env[63531]: DEBUG nova.network.neutron [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 531e1852-0c67-4b4d-a0ca-749ac438e79a] Updating instance_info_cache with network_info: [{"id": "99ca461b-afde-4ea5-8cbc-e596f9f8e528", "address": "fa:16:3e:94:cc:52", "network": {"id": "73e6d3ed-2016-498a-bda6-fd7dfbff93da", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1501868918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f727ddb992dc4c06954f25f6a1613b4a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c20f5114-0866-45b3-9a7c-62f113ff83fa", "external-id": "nsx-vlan-transportzone-47", "segmentation_id": 47, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap99ca461b-af", "ovs_interfaceid": "99ca461b-afde-4ea5-8cbc-e596f9f8e528", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.433316] env[63531]: DEBUG oslo_vmware.api [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]523f531e-acf7-719b-2e2e-db2f74cfb51b, 'name': SearchDatastore_Task, 'duration_secs': 0.009334} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.434844] env[63531]: DEBUG oslo_concurrency.lockutils [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.435362] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f/2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 807.435706] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c117a03b-3775-470a-a8a1-8be5fd67e37a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.443288] env[63531]: DEBUG oslo_vmware.api [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Waiting for the task: (returnval){ [ 807.443288] env[63531]: value = "task-1117804" [ 807.443288] env[63531]: _type = "Task" [ 807.443288] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.452111] env[63531]: DEBUG oslo_vmware.api [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': task-1117804, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.493115] env[63531]: DEBUG oslo_vmware.api [None req-ad241f4e-db19-4dc3-893a-2cb7737a96b1 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117802, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.722258] env[63531]: DEBUG oslo_vmware.api [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Task: {'id': task-1117803, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.785773] env[63531]: DEBUG oslo_vmware.rw_handles [None req-0acd5bb8-82ed-4cf7-9ea3-96d802548d79 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ac0d79-036e-727d-07c8-606921c098df/disk-0.vmdk. {{(pid=63531) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 807.785773] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b237c1f-484a-4c5a-b722-03a13627568d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.792112] env[63531]: DEBUG oslo_vmware.rw_handles [None req-0acd5bb8-82ed-4cf7-9ea3-96d802548d79 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ac0d79-036e-727d-07c8-606921c098df/disk-0.vmdk is in state: ready. {{(pid=63531) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 807.792291] env[63531]: ERROR oslo_vmware.rw_handles [None req-0acd5bb8-82ed-4cf7-9ea3-96d802548d79 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ac0d79-036e-727d-07c8-606921c098df/disk-0.vmdk due to incomplete transfer. [ 807.792511] env[63531]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-40e92e3e-a274-4bb9-b585-98276aee4179 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.804428] env[63531]: DEBUG oslo_vmware.rw_handles [None req-0acd5bb8-82ed-4cf7-9ea3-96d802548d79 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ac0d79-036e-727d-07c8-606921c098df/disk-0.vmdk. {{(pid=63531) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 807.804690] env[63531]: DEBUG nova.virt.vmwareapi.images [None req-0acd5bb8-82ed-4cf7-9ea3-96d802548d79 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Uploaded image 8197d52e-21f1-413d-8ef3-59307facd73d to the Glance image server {{(pid=63531) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 807.807505] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-0acd5bb8-82ed-4cf7-9ea3-96d802548d79 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Destroying the VM {{(pid=63531) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 807.807833] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-4ec5a2f7-0739-4cd6-a0a2-24f279933696 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.818325] env[63531]: DEBUG oslo_vmware.api [None req-0acd5bb8-82ed-4cf7-9ea3-96d802548d79 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Waiting for the task: (returnval){ [ 807.818325] env[63531]: value = "task-1117805" [ 807.818325] env[63531]: _type = "Task" [ 807.818325] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.830256] env[63531]: DEBUG oslo_vmware.api [None req-0acd5bb8-82ed-4cf7-9ea3-96d802548d79 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1117805, 'name': Destroy_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.900894] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Releasing lock "refresh_cache-531e1852-0c67-4b4d-a0ca-749ac438e79a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.901290] env[63531]: DEBUG nova.compute.manager [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 531e1852-0c67-4b4d-a0ca-749ac438e79a] Instance network_info: |[{"id": "99ca461b-afde-4ea5-8cbc-e596f9f8e528", "address": "fa:16:3e:94:cc:52", "network": {"id": "73e6d3ed-2016-498a-bda6-fd7dfbff93da", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1501868918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f727ddb992dc4c06954f25f6a1613b4a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c20f5114-0866-45b3-9a7c-62f113ff83fa", "external-id": "nsx-vlan-transportzone-47", "segmentation_id": 47, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap99ca461b-af", "ovs_interfaceid": "99ca461b-afde-4ea5-8cbc-e596f9f8e528", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 807.901806] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 531e1852-0c67-4b4d-a0ca-749ac438e79a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:94:cc:52', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c20f5114-0866-45b3-9a7c-62f113ff83fa', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '99ca461b-afde-4ea5-8cbc-e596f9f8e528', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 807.909931] env[63531]: DEBUG oslo.service.loopingcall [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 807.912899] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 531e1852-0c67-4b4d-a0ca-749ac438e79a] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 807.913381] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f6a2ab6c-bac8-4c4c-a36f-965ac8306538 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.938723] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 807.938723] env[63531]: value = "task-1117806" [ 807.938723] env[63531]: _type = "Task" [ 807.938723] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.955181] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117806, 'name': CreateVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.961895] env[63531]: DEBUG oslo_vmware.api [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': task-1117804, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.997256] env[63531]: DEBUG oslo_vmware.api [None req-ad241f4e-db19-4dc3-893a-2cb7737a96b1 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117802, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.222576] env[63531]: DEBUG oslo_vmware.api [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Task: {'id': task-1117803, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.248474] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b26c08b-2313-401e-8d9c-93651373e84b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.257530] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5488e5a4-7d19-43a1-a594-e3c2aa77fa1b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.296017] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26638d2d-0d57-47d5-a071-d55a10647cba {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.302833] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de70e615-3aa7-49b5-b41d-f6faad40dd4c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.317286] env[63531]: DEBUG nova.compute.provider_tree [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 808.328985] env[63531]: DEBUG oslo_vmware.api [None req-0acd5bb8-82ed-4cf7-9ea3-96d802548d79 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1117805, 'name': Destroy_Task, 'duration_secs': 0.383157} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.330304] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-0acd5bb8-82ed-4cf7-9ea3-96d802548d79 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Destroyed the VM [ 808.330634] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-0acd5bb8-82ed-4cf7-9ea3-96d802548d79 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Deleting Snapshot of the VM instance {{(pid=63531) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 808.330946] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-4231af56-5ae7-4c71-8792-c1db0ffb591f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.339377] env[63531]: DEBUG oslo_vmware.api [None req-0acd5bb8-82ed-4cf7-9ea3-96d802548d79 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Waiting for the task: (returnval){ [ 808.339377] env[63531]: value = "task-1117807" [ 808.339377] env[63531]: _type = "Task" [ 808.339377] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.347896] env[63531]: DEBUG oslo_vmware.api [None req-0acd5bb8-82ed-4cf7-9ea3-96d802548d79 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1117807, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.450514] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117806, 'name': CreateVM_Task} progress is 25%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.461449] env[63531]: DEBUG oslo_vmware.api [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': task-1117804, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.804984} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.461828] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f/2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 808.462177] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 808.462499] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-31adf6d7-0f65-4022-830e-7e020b734ec4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.470613] env[63531]: DEBUG oslo_vmware.api [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Waiting for the task: (returnval){ [ 808.470613] env[63531]: value = "task-1117808" [ 808.470613] env[63531]: _type = "Task" [ 808.470613] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.479588] env[63531]: DEBUG oslo_vmware.api [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': task-1117808, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.493033] env[63531]: DEBUG oslo_vmware.api [None req-ad241f4e-db19-4dc3-893a-2cb7737a96b1 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117802, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.723300] env[63531]: DEBUG oslo_vmware.api [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Task: {'id': task-1117803, 'name': PowerOnVM_Task, 'duration_secs': 1.410458} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.723565] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 808.723775] env[63531]: INFO nova.compute.manager [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Took 9.15 seconds to spawn the instance on the hypervisor. [ 808.723963] env[63531]: DEBUG nova.compute.manager [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 808.724905] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32b80c1f-b4d9-4610-8b23-31df48e592fe {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.826159] env[63531]: DEBUG nova.scheduler.client.report [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 808.850916] env[63531]: DEBUG oslo_vmware.api [None req-0acd5bb8-82ed-4cf7-9ea3-96d802548d79 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1117807, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.951146] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117806, 'name': CreateVM_Task, 'duration_secs': 0.925791} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.951474] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 531e1852-0c67-4b4d-a0ca-749ac438e79a] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 808.952079] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.952237] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.952573] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 808.952824] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-43394b42-af14-4b9c-bbcf-8bd10cea7bcb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.957517] env[63531]: DEBUG oslo_vmware.api [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Waiting for the task: (returnval){ [ 808.957517] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52063572-9f18-d11a-39d0-d8ab502cbbe3" [ 808.957517] env[63531]: _type = "Task" [ 808.957517] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.965539] env[63531]: DEBUG oslo_vmware.api [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52063572-9f18-d11a-39d0-d8ab502cbbe3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.981173] env[63531]: DEBUG oslo_vmware.api [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': task-1117808, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063155} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.981451] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 808.982236] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b65b4623-3628-4b53-b5df-5fedb7429cc1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.008037] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Reconfiguring VM instance instance-00000039 to attach disk [datastore1] 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f/2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 809.012266] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5ef0fa0b-7b37-47fe-a9c5-a05fd511c69a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.026781] env[63531]: DEBUG oslo_vmware.api [None req-ad241f4e-db19-4dc3-893a-2cb7737a96b1 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117802, 'name': CreateSnapshot_Task, 'duration_secs': 1.914893} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.027106] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-ad241f4e-db19-4dc3-893a-2cb7737a96b1 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Created Snapshot of the VM instance {{(pid=63531) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 809.028269] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-309ab349-0b74-4f90-bd5d-46dac556339b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.032368] env[63531]: DEBUG oslo_vmware.api [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Waiting for the task: (returnval){ [ 809.032368] env[63531]: value = "task-1117809" [ 809.032368] env[63531]: _type = "Task" [ 809.032368] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.045775] env[63531]: DEBUG oslo_vmware.api [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': task-1117809, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.251021] env[63531]: INFO nova.compute.manager [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Took 39.30 seconds to build instance. [ 809.331648] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.550s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.332205] env[63531]: DEBUG nova.compute.manager [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] [instance: eb6c84d7-e0e6-4999-9018-889b99dc67ad] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 809.335174] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.139s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.336658] env[63531]: INFO nova.compute.claims [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 809.348964] env[63531]: DEBUG oslo_vmware.api [None req-0acd5bb8-82ed-4cf7-9ea3-96d802548d79 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1117807, 'name': RemoveSnapshot_Task, 'duration_secs': 0.978226} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.349301] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-0acd5bb8-82ed-4cf7-9ea3-96d802548d79 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Deleted Snapshot of the VM instance {{(pid=63531) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 809.349538] env[63531]: INFO nova.compute.manager [None req-0acd5bb8-82ed-4cf7-9ea3-96d802548d79 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Took 14.84 seconds to snapshot the instance on the hypervisor. [ 809.473197] env[63531]: DEBUG oslo_vmware.api [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52063572-9f18-d11a-39d0-d8ab502cbbe3, 'name': SearchDatastore_Task, 'duration_secs': 0.044095} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.473197] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.473790] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 531e1852-0c67-4b4d-a0ca-749ac438e79a] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 809.473790] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.473790] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.473950] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 809.474130] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-44303784-11ec-4daa-acc9-32c68fb108e6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.484535] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 809.484736] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 809.485535] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bb93f81f-5f28-425d-8da9-26394ccb96b3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.491308] env[63531]: DEBUG oslo_vmware.api [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Waiting for the task: (returnval){ [ 809.491308] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52953770-cf18-29bb-b22b-ecd1828d8161" [ 809.491308] env[63531]: _type = "Task" [ 809.491308] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.499182] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Acquiring lock "92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.499861] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Lock "92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.504094] env[63531]: DEBUG oslo_vmware.api [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52953770-cf18-29bb-b22b-ecd1828d8161, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.556230] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-ad241f4e-db19-4dc3-893a-2cb7737a96b1 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Creating linked-clone VM from snapshot {{(pid=63531) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 809.558317] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-e643e994-3dbf-454b-ae3b-08e886e15993 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.571911] env[63531]: DEBUG oslo_vmware.api [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': task-1117809, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.575843] env[63531]: DEBUG oslo_vmware.api [None req-ad241f4e-db19-4dc3-893a-2cb7737a96b1 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Waiting for the task: (returnval){ [ 809.575843] env[63531]: value = "task-1117810" [ 809.575843] env[63531]: _type = "Task" [ 809.575843] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.585624] env[63531]: DEBUG oslo_vmware.api [None req-ad241f4e-db19-4dc3-893a-2cb7737a96b1 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117810, 'name': CloneVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.750535] env[63531]: DEBUG oslo_concurrency.lockutils [None req-09ffe008-9d41-475a-bb2e-bafb1cadc429 tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Lock "caf0dee8-cb5a-402b-b533-22a590433f54" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 100.108s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.846111] env[63531]: DEBUG nova.compute.utils [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 809.848910] env[63531]: DEBUG nova.compute.manager [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] [instance: eb6c84d7-e0e6-4999-9018-889b99dc67ad] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 809.849311] env[63531]: DEBUG nova.network.neutron [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] [instance: eb6c84d7-e0e6-4999-9018-889b99dc67ad] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 809.933543] env[63531]: DEBUG nova.policy [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '57a8e0522c284b1c96798ca6c0bf3878', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1093a53d17004ed0984d01aed7487993', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 810.005815] env[63531]: DEBUG oslo_vmware.api [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52953770-cf18-29bb-b22b-ecd1828d8161, 'name': SearchDatastore_Task, 'duration_secs': 0.018547} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.007373] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f767afcc-dbc0-49ca-81eb-42d752aad9d3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.013031] env[63531]: DEBUG oslo_vmware.api [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Waiting for the task: (returnval){ [ 810.013031] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]520a5b1f-2ceb-6629-4b27-500e683e7f3f" [ 810.013031] env[63531]: _type = "Task" [ 810.013031] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.021509] env[63531]: DEBUG oslo_vmware.api [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]520a5b1f-2ceb-6629-4b27-500e683e7f3f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.047075] env[63531]: DEBUG oslo_vmware.api [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': task-1117809, 'name': ReconfigVM_Task, 'duration_secs': 0.919835} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.047394] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Reconfigured VM instance instance-00000039 to attach disk [datastore1] 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f/2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 810.048057] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ee793280-ab99-49dc-890a-b3958beea1b5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.055139] env[63531]: DEBUG nova.compute.manager [req-e7bfce74-6e1c-40e8-8c49-25b93d447101 req-89e72732-eca9-4f8e-ac46-4d11c8ec4fa4 service nova] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Received event network-changed-d99cda7f-476b-4869-8ba7-4a00c03da394 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 810.055318] env[63531]: DEBUG nova.compute.manager [req-e7bfce74-6e1c-40e8-8c49-25b93d447101 req-89e72732-eca9-4f8e-ac46-4d11c8ec4fa4 service nova] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Refreshing instance network info cache due to event network-changed-d99cda7f-476b-4869-8ba7-4a00c03da394. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 810.055541] env[63531]: DEBUG oslo_concurrency.lockutils [req-e7bfce74-6e1c-40e8-8c49-25b93d447101 req-89e72732-eca9-4f8e-ac46-4d11c8ec4fa4 service nova] Acquiring lock "refresh_cache-caf0dee8-cb5a-402b-b533-22a590433f54" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 810.055697] env[63531]: DEBUG oslo_concurrency.lockutils [req-e7bfce74-6e1c-40e8-8c49-25b93d447101 req-89e72732-eca9-4f8e-ac46-4d11c8ec4fa4 service nova] Acquired lock "refresh_cache-caf0dee8-cb5a-402b-b533-22a590433f54" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.055862] env[63531]: DEBUG nova.network.neutron [req-e7bfce74-6e1c-40e8-8c49-25b93d447101 req-89e72732-eca9-4f8e-ac46-4d11c8ec4fa4 service nova] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Refreshing network info cache for port d99cda7f-476b-4869-8ba7-4a00c03da394 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 810.059131] env[63531]: DEBUG oslo_vmware.api [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Waiting for the task: (returnval){ [ 810.059131] env[63531]: value = "task-1117811" [ 810.059131] env[63531]: _type = "Task" [ 810.059131] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.071078] env[63531]: DEBUG oslo_vmware.api [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': task-1117811, 'name': Rename_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.085133] env[63531]: DEBUG oslo_vmware.api [None req-ad241f4e-db19-4dc3-893a-2cb7737a96b1 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117810, 'name': CloneVM_Task} progress is 94%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.253681] env[63531]: DEBUG nova.compute.manager [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 810.356042] env[63531]: DEBUG nova.compute.manager [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] [instance: eb6c84d7-e0e6-4999-9018-889b99dc67ad] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 810.360227] env[63531]: DEBUG nova.network.neutron [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] [instance: eb6c84d7-e0e6-4999-9018-889b99dc67ad] Successfully created port: 01bd4815-38ea-44f3-b7d3-c24fef3fd863 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 810.423459] env[63531]: DEBUG oslo_concurrency.lockutils [None req-28bc9bb5-b018-4dad-a1e2-883737c98fb3 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "29dd6de3-2765-408c-acc0-da47e5e0a977" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.423815] env[63531]: DEBUG oslo_concurrency.lockutils [None req-28bc9bb5-b018-4dad-a1e2-883737c98fb3 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "29dd6de3-2765-408c-acc0-da47e5e0a977" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 810.424017] env[63531]: DEBUG nova.compute.manager [None req-28bc9bb5-b018-4dad-a1e2-883737c98fb3 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 810.425381] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa127a5f-2f18-4359-87fa-2e5d89eb6744 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.439273] env[63531]: DEBUG nova.compute.manager [None req-28bc9bb5-b018-4dad-a1e2-883737c98fb3 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63531) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 810.439788] env[63531]: DEBUG nova.objects.instance [None req-28bc9bb5-b018-4dad-a1e2-883737c98fb3 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lazy-loading 'flavor' on Instance uuid 29dd6de3-2765-408c-acc0-da47e5e0a977 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 810.525585] env[63531]: DEBUG oslo_vmware.api [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]520a5b1f-2ceb-6629-4b27-500e683e7f3f, 'name': SearchDatastore_Task, 'duration_secs': 0.021914} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.528699] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 810.528699] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] 531e1852-0c67-4b4d-a0ca-749ac438e79a/531e1852-0c67-4b4d-a0ca-749ac438e79a.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 810.529594] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e1e9e162-27bc-4046-b48c-f7922df9871d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.539105] env[63531]: DEBUG oslo_vmware.api [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Waiting for the task: (returnval){ [ 810.539105] env[63531]: value = "task-1117812" [ 810.539105] env[63531]: _type = "Task" [ 810.539105] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.551520] env[63531]: DEBUG oslo_vmware.api [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1117812, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.569608] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Acquiring lock "db83adf4-6183-4330-b260-77d1f5daf899" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.569839] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Lock "db83adf4-6183-4330-b260-77d1f5daf899" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 810.576243] env[63531]: DEBUG oslo_vmware.api [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': task-1117811, 'name': Rename_Task, 'duration_secs': 0.318896} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.582641] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 810.583881] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-52060a52-7621-408d-8e21-75b3971cbe3c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.590965] env[63531]: DEBUG oslo_vmware.api [None req-ad241f4e-db19-4dc3-893a-2cb7737a96b1 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117810, 'name': CloneVM_Task} progress is 94%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.595885] env[63531]: DEBUG oslo_vmware.api [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Waiting for the task: (returnval){ [ 810.595885] env[63531]: value = "task-1117813" [ 810.595885] env[63531]: _type = "Task" [ 810.595885] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.601989] env[63531]: DEBUG oslo_vmware.api [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': task-1117813, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.764384] env[63531]: DEBUG oslo_concurrency.lockutils [None req-decfc0cf-abaa-40f8-a951-07480044abae tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Acquiring lock "caf0dee8-cb5a-402b-b533-22a590433f54" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.764665] env[63531]: DEBUG oslo_concurrency.lockutils [None req-decfc0cf-abaa-40f8-a951-07480044abae tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Lock "caf0dee8-cb5a-402b-b533-22a590433f54" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 810.764980] env[63531]: INFO nova.compute.manager [None req-decfc0cf-abaa-40f8-a951-07480044abae tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Rebooting instance [ 810.794791] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.847974] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5129d871-9d9a-4436-9375-99ef1887bb7d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.855894] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-389812cf-3830-4e06-b272-9d7132a21a2b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.896361] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2eb5c95c-659a-448b-bd29-8465ff01f680 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.904649] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c88dc311-db53-418f-99ac-c09be28c3121 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.919872] env[63531]: DEBUG nova.compute.provider_tree [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 810.945859] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-28bc9bb5-b018-4dad-a1e2-883737c98fb3 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 810.946301] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-be03a5dd-f6b4-4951-af2d-4d70b05608af {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.954698] env[63531]: DEBUG oslo_vmware.api [None req-28bc9bb5-b018-4dad-a1e2-883737c98fb3 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 810.954698] env[63531]: value = "task-1117814" [ 810.954698] env[63531]: _type = "Task" [ 810.954698] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.964451] env[63531]: DEBUG oslo_vmware.api [None req-28bc9bb5-b018-4dad-a1e2-883737c98fb3 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1117814, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.973167] env[63531]: DEBUG nova.network.neutron [req-e7bfce74-6e1c-40e8-8c49-25b93d447101 req-89e72732-eca9-4f8e-ac46-4d11c8ec4fa4 service nova] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Updated VIF entry in instance network info cache for port d99cda7f-476b-4869-8ba7-4a00c03da394. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 810.973596] env[63531]: DEBUG nova.network.neutron [req-e7bfce74-6e1c-40e8-8c49-25b93d447101 req-89e72732-eca9-4f8e-ac46-4d11c8ec4fa4 service nova] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Updating instance_info_cache with network_info: [{"id": "d99cda7f-476b-4869-8ba7-4a00c03da394", "address": "fa:16:3e:9d:85:ab", "network": {"id": "067f0236-aa59-4a63-a33c-455dc28e0c1a", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-54491452-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a0e384d6cf184a98aa3602eca6b3f410", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7bdf594e-da7a-4254-b413-87aef4614588", "external-id": "nsx-vlan-transportzone-422", "segmentation_id": 422, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd99cda7f-47", "ovs_interfaceid": "d99cda7f-476b-4869-8ba7-4a00c03da394", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 811.050079] env[63531]: DEBUG oslo_vmware.api [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1117812, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.085718] env[63531]: DEBUG oslo_vmware.api [None req-ad241f4e-db19-4dc3-893a-2cb7737a96b1 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117810, 'name': CloneVM_Task} progress is 100%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.102472] env[63531]: DEBUG oslo_vmware.api [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': task-1117813, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.306517] env[63531]: DEBUG oslo_concurrency.lockutils [None req-decfc0cf-abaa-40f8-a951-07480044abae tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Acquiring lock "refresh_cache-caf0dee8-cb5a-402b-b533-22a590433f54" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.371184] env[63531]: DEBUG nova.compute.manager [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] [instance: eb6c84d7-e0e6-4999-9018-889b99dc67ad] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 811.396186] env[63531]: DEBUG nova.virt.hardware [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 811.396500] env[63531]: DEBUG nova.virt.hardware [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 811.396698] env[63531]: DEBUG nova.virt.hardware [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 811.396904] env[63531]: DEBUG nova.virt.hardware [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 811.397150] env[63531]: DEBUG nova.virt.hardware [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 811.397332] env[63531]: DEBUG nova.virt.hardware [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 811.397569] env[63531]: DEBUG nova.virt.hardware [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 811.397749] env[63531]: DEBUG nova.virt.hardware [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 811.397927] env[63531]: DEBUG nova.virt.hardware [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 811.398165] env[63531]: DEBUG nova.virt.hardware [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 811.398369] env[63531]: DEBUG nova.virt.hardware [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 811.399337] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a5bd38a-6f73-41b0-9297-7f439249e314 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.407874] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abb8e7a9-35e4-4470-9c00-a8211eafbbeb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.423168] env[63531]: DEBUG nova.scheduler.client.report [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 811.465152] env[63531]: DEBUG oslo_vmware.api [None req-28bc9bb5-b018-4dad-a1e2-883737c98fb3 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1117814, 'name': PowerOffVM_Task, 'duration_secs': 0.215398} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.465463] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-28bc9bb5-b018-4dad-a1e2-883737c98fb3 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 811.465660] env[63531]: DEBUG nova.compute.manager [None req-28bc9bb5-b018-4dad-a1e2-883737c98fb3 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 811.466515] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6feafeb2-88a3-45dd-88c0-d34c046a0fee {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.477595] env[63531]: DEBUG oslo_concurrency.lockutils [req-e7bfce74-6e1c-40e8-8c49-25b93d447101 req-89e72732-eca9-4f8e-ac46-4d11c8ec4fa4 service nova] Releasing lock "refresh_cache-caf0dee8-cb5a-402b-b533-22a590433f54" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 811.477595] env[63531]: DEBUG oslo_concurrency.lockutils [None req-decfc0cf-abaa-40f8-a951-07480044abae tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Acquired lock "refresh_cache-caf0dee8-cb5a-402b-b533-22a590433f54" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.477595] env[63531]: DEBUG nova.network.neutron [None req-decfc0cf-abaa-40f8-a951-07480044abae tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 811.550731] env[63531]: DEBUG oslo_vmware.api [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1117812, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.552096} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.551050] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] 531e1852-0c67-4b4d-a0ca-749ac438e79a/531e1852-0c67-4b4d-a0ca-749ac438e79a.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 811.551255] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 531e1852-0c67-4b4d-a0ca-749ac438e79a] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 811.551522] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8d458d4c-0014-425a-91cc-452d8a898a29 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.562166] env[63531]: DEBUG oslo_vmware.api [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Waiting for the task: (returnval){ [ 811.562166] env[63531]: value = "task-1117815" [ 811.562166] env[63531]: _type = "Task" [ 811.562166] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.568577] env[63531]: DEBUG oslo_vmware.api [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1117815, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.586603] env[63531]: DEBUG oslo_vmware.api [None req-ad241f4e-db19-4dc3-893a-2cb7737a96b1 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117810, 'name': CloneVM_Task, 'duration_secs': 1.642607} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.589232] env[63531]: INFO nova.virt.vmwareapi.vmops [None req-ad241f4e-db19-4dc3-893a-2cb7737a96b1 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Created linked-clone VM from snapshot [ 811.589232] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ede8e12f-4dad-4760-b780-84d3c91468bd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.597676] env[63531]: DEBUG nova.virt.vmwareapi.images [None req-ad241f4e-db19-4dc3-893a-2cb7737a96b1 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Uploading image 2a1376af-2b52-4754-afef-6403cc9e22a3 {{(pid=63531) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 811.607070] env[63531]: DEBUG oslo_vmware.api [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': task-1117813, 'name': PowerOnVM_Task, 'duration_secs': 0.653392} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.607359] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 811.607587] env[63531]: INFO nova.compute.manager [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Took 9.51 seconds to spawn the instance on the hypervisor. [ 811.607793] env[63531]: DEBUG nova.compute.manager [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 811.610408] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61dac6d7-d428-4290-8eb8-01661210db62 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.631809] env[63531]: DEBUG oslo_vmware.rw_handles [None req-ad241f4e-db19-4dc3-893a-2cb7737a96b1 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 811.631809] env[63531]: value = "vm-244680" [ 811.631809] env[63531]: _type = "VirtualMachine" [ 811.631809] env[63531]: }. {{(pid=63531) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 811.632181] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-e25bfd0d-b67a-4330-803a-bf64cb14455a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.640597] env[63531]: DEBUG oslo_vmware.rw_handles [None req-ad241f4e-db19-4dc3-893a-2cb7737a96b1 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Lease: (returnval){ [ 811.640597] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5202c88e-efd1-54c2-7926-2cc42a2bbebe" [ 811.640597] env[63531]: _type = "HttpNfcLease" [ 811.640597] env[63531]: } obtained for exporting VM: (result){ [ 811.640597] env[63531]: value = "vm-244680" [ 811.640597] env[63531]: _type = "VirtualMachine" [ 811.640597] env[63531]: }. {{(pid=63531) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 811.640841] env[63531]: DEBUG oslo_vmware.api [None req-ad241f4e-db19-4dc3-893a-2cb7737a96b1 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Waiting for the lease: (returnval){ [ 811.640841] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5202c88e-efd1-54c2-7926-2cc42a2bbebe" [ 811.640841] env[63531]: _type = "HttpNfcLease" [ 811.640841] env[63531]: } to be ready. {{(pid=63531) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 811.647653] env[63531]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 811.647653] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5202c88e-efd1-54c2-7926-2cc42a2bbebe" [ 811.647653] env[63531]: _type = "HttpNfcLease" [ 811.647653] env[63531]: } is initializing. {{(pid=63531) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 811.928671] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.593s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.929577] env[63531]: DEBUG nova.compute.manager [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 811.933678] env[63531]: DEBUG oslo_concurrency.lockutils [None req-637af179-656f-4b75-bdcc-db626dfdd5bd tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.560s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 811.986225] env[63531]: DEBUG oslo_concurrency.lockutils [None req-28bc9bb5-b018-4dad-a1e2-883737c98fb3 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "29dd6de3-2765-408c-acc0-da47e5e0a977" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.562s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.076114] env[63531]: DEBUG oslo_vmware.api [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1117815, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.305134} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.082034] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 531e1852-0c67-4b4d-a0ca-749ac438e79a] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 812.082034] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b045c35-4ef1-48b3-b3c5-93b31fb36fe4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.106082] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 531e1852-0c67-4b4d-a0ca-749ac438e79a] Reconfiguring VM instance instance-0000003a to attach disk [datastore2] 531e1852-0c67-4b4d-a0ca-749ac438e79a/531e1852-0c67-4b4d-a0ca-749ac438e79a.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 812.112206] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0ec20e1a-69a9-43db-86fc-286591197f17 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.127713] env[63531]: DEBUG nova.compute.manager [req-782a32c5-2b6e-48a9-8824-a5355ef92488 req-4b0d21bf-bbd0-4c85-8c62-55cd4c697ff6 service nova] [instance: eb6c84d7-e0e6-4999-9018-889b99dc67ad] Received event network-vif-plugged-01bd4815-38ea-44f3-b7d3-c24fef3fd863 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 812.127947] env[63531]: DEBUG oslo_concurrency.lockutils [req-782a32c5-2b6e-48a9-8824-a5355ef92488 req-4b0d21bf-bbd0-4c85-8c62-55cd4c697ff6 service nova] Acquiring lock "eb6c84d7-e0e6-4999-9018-889b99dc67ad-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 812.128209] env[63531]: DEBUG oslo_concurrency.lockutils [req-782a32c5-2b6e-48a9-8824-a5355ef92488 req-4b0d21bf-bbd0-4c85-8c62-55cd4c697ff6 service nova] Lock "eb6c84d7-e0e6-4999-9018-889b99dc67ad-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 812.128429] env[63531]: DEBUG oslo_concurrency.lockutils [req-782a32c5-2b6e-48a9-8824-a5355ef92488 req-4b0d21bf-bbd0-4c85-8c62-55cd4c697ff6 service nova] Lock "eb6c84d7-e0e6-4999-9018-889b99dc67ad-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.128854] env[63531]: DEBUG nova.compute.manager [req-782a32c5-2b6e-48a9-8824-a5355ef92488 req-4b0d21bf-bbd0-4c85-8c62-55cd4c697ff6 service nova] [instance: eb6c84d7-e0e6-4999-9018-889b99dc67ad] No waiting events found dispatching network-vif-plugged-01bd4815-38ea-44f3-b7d3-c24fef3fd863 {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 812.129088] env[63531]: WARNING nova.compute.manager [req-782a32c5-2b6e-48a9-8824-a5355ef92488 req-4b0d21bf-bbd0-4c85-8c62-55cd4c697ff6 service nova] [instance: eb6c84d7-e0e6-4999-9018-889b99dc67ad] Received unexpected event network-vif-plugged-01bd4815-38ea-44f3-b7d3-c24fef3fd863 for instance with vm_state building and task_state spawning. [ 812.142389] env[63531]: DEBUG oslo_vmware.api [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Waiting for the task: (returnval){ [ 812.142389] env[63531]: value = "task-1117817" [ 812.142389] env[63531]: _type = "Task" [ 812.142389] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.142389] env[63531]: INFO nova.compute.manager [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Took 40.68 seconds to build instance. [ 812.161234] env[63531]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 812.161234] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5202c88e-efd1-54c2-7926-2cc42a2bbebe" [ 812.161234] env[63531]: _type = "HttpNfcLease" [ 812.161234] env[63531]: } is ready. {{(pid=63531) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 812.167660] env[63531]: DEBUG oslo_vmware.rw_handles [None req-ad241f4e-db19-4dc3-893a-2cb7737a96b1 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 812.167660] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5202c88e-efd1-54c2-7926-2cc42a2bbebe" [ 812.167660] env[63531]: _type = "HttpNfcLease" [ 812.167660] env[63531]: }. {{(pid=63531) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 812.169238] env[63531]: DEBUG oslo_vmware.api [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1117817, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.169713] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-989c27d0-1645-4d5f-96b8-0ed7f0bef585 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.183062] env[63531]: DEBUG oslo_vmware.rw_handles [None req-ad241f4e-db19-4dc3-893a-2cb7737a96b1 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5210c391-01c5-4cde-8230-d9ed42991e74/disk-0.vmdk from lease info. {{(pid=63531) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 812.183062] env[63531]: DEBUG oslo_vmware.rw_handles [None req-ad241f4e-db19-4dc3-893a-2cb7737a96b1 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5210c391-01c5-4cde-8230-d9ed42991e74/disk-0.vmdk for reading. {{(pid=63531) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 812.268246] env[63531]: DEBUG nova.network.neutron [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] [instance: eb6c84d7-e0e6-4999-9018-889b99dc67ad] Successfully updated port: 01bd4815-38ea-44f3-b7d3-c24fef3fd863 {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 812.327331] env[63531]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-6d5dc2b3-fd8d-4f18-822c-14629f69d95c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.422611] env[63531]: DEBUG nova.network.neutron [None req-decfc0cf-abaa-40f8-a951-07480044abae tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Updating instance_info_cache with network_info: [{"id": "d99cda7f-476b-4869-8ba7-4a00c03da394", "address": "fa:16:3e:9d:85:ab", "network": {"id": "067f0236-aa59-4a63-a33c-455dc28e0c1a", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-54491452-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a0e384d6cf184a98aa3602eca6b3f410", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7bdf594e-da7a-4254-b413-87aef4614588", "external-id": "nsx-vlan-transportzone-422", "segmentation_id": 422, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd99cda7f-47", "ovs_interfaceid": "d99cda7f-476b-4869-8ba7-4a00c03da394", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 812.438902] env[63531]: DEBUG nova.compute.utils [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 812.440590] env[63531]: DEBUG nova.compute.manager [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 812.440761] env[63531]: DEBUG nova.network.neutron [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 812.503780] env[63531]: DEBUG nova.policy [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ff7b1cf5f24e47cba315f4c4ee47e131', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ee5fd84e1da04fd984f81578e4085850', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 812.506979] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b92df4e-58dc-48e8-8526-1b04ea84cbdc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.514608] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-634c9764-41bc-47c1-82f8-e11b34719e48 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.546511] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfa77cc3-a8ec-4906-821c-0e45975d7e43 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.554812] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe367ef7-ef4f-4915-bce6-6f69f06df146 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.571382] env[63531]: DEBUG nova.compute.provider_tree [None req-637af179-656f-4b75-bdcc-db626dfdd5bd tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 812.649261] env[63531]: DEBUG oslo_concurrency.lockutils [None req-78a9451a-f13b-487e-8619-df07dab4c9fd tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Lock "2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 96.555s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.656944] env[63531]: DEBUG oslo_vmware.api [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1117817, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.773171] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Acquiring lock "refresh_cache-eb6c84d7-e0e6-4999-9018-889b99dc67ad" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.773334] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Acquired lock "refresh_cache-eb6c84d7-e0e6-4999-9018-889b99dc67ad" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.773473] env[63531]: DEBUG nova.network.neutron [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] [instance: eb6c84d7-e0e6-4999-9018-889b99dc67ad] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 812.927878] env[63531]: DEBUG oslo_concurrency.lockutils [None req-decfc0cf-abaa-40f8-a951-07480044abae tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Releasing lock "refresh_cache-caf0dee8-cb5a-402b-b533-22a590433f54" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.930019] env[63531]: DEBUG nova.compute.manager [None req-decfc0cf-abaa-40f8-a951-07480044abae tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 812.932065] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-314503c0-6c02-4e1e-bb7e-97175e47b538 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.944313] env[63531]: DEBUG nova.compute.manager [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 813.074992] env[63531]: DEBUG nova.scheduler.client.report [None req-637af179-656f-4b75-bdcc-db626dfdd5bd tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 813.139781] env[63531]: DEBUG nova.network.neutron [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Successfully created port: a3fd640b-c47b-4501-ac04-f90f6af8b076 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 813.155863] env[63531]: DEBUG nova.compute.manager [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 813.168655] env[63531]: DEBUG oslo_vmware.api [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1117817, 'name': ReconfigVM_Task, 'duration_secs': 0.707159} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.168988] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 531e1852-0c67-4b4d-a0ca-749ac438e79a] Reconfigured VM instance instance-0000003a to attach disk [datastore2] 531e1852-0c67-4b4d-a0ca-749ac438e79a/531e1852-0c67-4b4d-a0ca-749ac438e79a.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 813.169694] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-03306d98-2c2b-4225-8eef-73b9174f80f5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.177536] env[63531]: DEBUG oslo_vmware.api [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Waiting for the task: (returnval){ [ 813.177536] env[63531]: value = "task-1117818" [ 813.177536] env[63531]: _type = "Task" [ 813.177536] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.189709] env[63531]: DEBUG oslo_vmware.api [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1117818, 'name': Rename_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.455648] env[63531]: DEBUG nova.network.neutron [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] [instance: eb6c84d7-e0e6-4999-9018-889b99dc67ad] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 813.581670] env[63531]: DEBUG oslo_concurrency.lockutils [None req-637af179-656f-4b75-bdcc-db626dfdd5bd tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.648s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 813.581948] env[63531]: INFO nova.compute.manager [None req-637af179-656f-4b75-bdcc-db626dfdd5bd tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Successfully reverted task state from rebuilding on failure for instance. [ 813.589868] env[63531]: DEBUG oslo_concurrency.lockutils [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.019s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.591510] env[63531]: INFO nova.compute.claims [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 813.648049] env[63531]: DEBUG nova.objects.instance [None req-bec51a9c-2a9e-4d84-8fff-a20a798f4290 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lazy-loading 'flavor' on Instance uuid 29dd6de3-2765-408c-acc0-da47e5e0a977 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 813.683633] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.690794] env[63531]: DEBUG oslo_vmware.api [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1117818, 'name': Rename_Task, 'duration_secs': 0.229129} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.691179] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 531e1852-0c67-4b4d-a0ca-749ac438e79a] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 813.691465] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6f7de24d-50ee-40ac-b273-8c9487d27feb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.698186] env[63531]: DEBUG oslo_vmware.api [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Waiting for the task: (returnval){ [ 813.698186] env[63531]: value = "task-1117819" [ 813.698186] env[63531]: _type = "Task" [ 813.698186] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.704234] env[63531]: DEBUG nova.network.neutron [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] [instance: eb6c84d7-e0e6-4999-9018-889b99dc67ad] Updating instance_info_cache with network_info: [{"id": "01bd4815-38ea-44f3-b7d3-c24fef3fd863", "address": "fa:16:3e:50:3e:97", "network": {"id": "c818e149-d1f0-4573-a514-996cb03d2e96", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-372309638-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1093a53d17004ed0984d01aed7487993", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1e029825-6c65-4ac7-88f6-65f9d106db76", "external-id": "nsx-vlan-transportzone-428", "segmentation_id": 428, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01bd4815-38", "ovs_interfaceid": "01bd4815-38ea-44f3-b7d3-c24fef3fd863", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 813.711857] env[63531]: DEBUG oslo_vmware.api [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1117819, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.951129] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-013025b4-eeb1-4fc3-9c86-16d7af841669 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.958873] env[63531]: DEBUG nova.compute.manager [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 813.968559] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-decfc0cf-abaa-40f8-a951-07480044abae tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Doing hard reboot of VM {{(pid=63531) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1064}} [ 813.970022] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-9e5aee3b-1a9b-432d-98f2-f81e0b53809a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.978889] env[63531]: DEBUG oslo_vmware.api [None req-decfc0cf-abaa-40f8-a951-07480044abae tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Waiting for the task: (returnval){ [ 813.978889] env[63531]: value = "task-1117820" [ 813.978889] env[63531]: _type = "Task" [ 813.978889] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.999777] env[63531]: DEBUG oslo_vmware.api [None req-decfc0cf-abaa-40f8-a951-07480044abae tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Task: {'id': task-1117820, 'name': ResetVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.012965] env[63531]: DEBUG nova.virt.hardware [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 814.013656] env[63531]: DEBUG nova.virt.hardware [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 814.013805] env[63531]: DEBUG nova.virt.hardware [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 814.014029] env[63531]: DEBUG nova.virt.hardware [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 814.014203] env[63531]: DEBUG nova.virt.hardware [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 814.014371] env[63531]: DEBUG nova.virt.hardware [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 814.014592] env[63531]: DEBUG nova.virt.hardware [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 814.014907] env[63531]: DEBUG nova.virt.hardware [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 814.015086] env[63531]: DEBUG nova.virt.hardware [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 814.015791] env[63531]: DEBUG nova.virt.hardware [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 814.015791] env[63531]: DEBUG nova.virt.hardware [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 814.016707] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-548fba68-5b55-4b07-ba05-ee5fda98aae7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.026758] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3efaed68-79d1-4b80-8978-d50a6813c81e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.155932] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bec51a9c-2a9e-4d84-8fff-a20a798f4290 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "refresh_cache-29dd6de3-2765-408c-acc0-da47e5e0a977" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.156781] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bec51a9c-2a9e-4d84-8fff-a20a798f4290 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquired lock "refresh_cache-29dd6de3-2765-408c-acc0-da47e5e0a977" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.158170] env[63531]: DEBUG nova.network.neutron [None req-bec51a9c-2a9e-4d84-8fff-a20a798f4290 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 814.158543] env[63531]: DEBUG nova.objects.instance [None req-bec51a9c-2a9e-4d84-8fff-a20a798f4290 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lazy-loading 'info_cache' on Instance uuid 29dd6de3-2765-408c-acc0-da47e5e0a977 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 814.211710] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Releasing lock "refresh_cache-eb6c84d7-e0e6-4999-9018-889b99dc67ad" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 814.212365] env[63531]: DEBUG nova.compute.manager [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] [instance: eb6c84d7-e0e6-4999-9018-889b99dc67ad] Instance network_info: |[{"id": "01bd4815-38ea-44f3-b7d3-c24fef3fd863", "address": "fa:16:3e:50:3e:97", "network": {"id": "c818e149-d1f0-4573-a514-996cb03d2e96", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-372309638-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1093a53d17004ed0984d01aed7487993", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1e029825-6c65-4ac7-88f6-65f9d106db76", "external-id": "nsx-vlan-transportzone-428", "segmentation_id": 428, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01bd4815-38", "ovs_interfaceid": "01bd4815-38ea-44f3-b7d3-c24fef3fd863", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 814.212902] env[63531]: DEBUG oslo_vmware.api [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1117819, 'name': PowerOnVM_Task} progress is 92%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.213698] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] [instance: eb6c84d7-e0e6-4999-9018-889b99dc67ad] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:50:3e:97', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1e029825-6c65-4ac7-88f6-65f9d106db76', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '01bd4815-38ea-44f3-b7d3-c24fef3fd863', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 814.222968] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Creating folder: Project (1093a53d17004ed0984d01aed7487993). Parent ref: group-v244585. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 814.223473] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-af78fb31-90a3-4bc3-9858-ae34b9cbc3dd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.236528] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Created folder: Project (1093a53d17004ed0984d01aed7487993) in parent group-v244585. [ 814.237118] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Creating folder: Instances. Parent ref: group-v244681. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 814.237529] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f6859eab-46ad-4390-83a4-0ad30ee55a06 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.247533] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Created folder: Instances in parent group-v244681. [ 814.248521] env[63531]: DEBUG oslo.service.loopingcall [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 814.248884] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eb6c84d7-e0e6-4999-9018-889b99dc67ad] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 814.249227] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5f0c7b75-869a-4f2e-a4f8-09c8991ca920 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.270617] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 814.270617] env[63531]: value = "task-1117823" [ 814.270617] env[63531]: _type = "Task" [ 814.270617] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.279818] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117823, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.491394] env[63531]: DEBUG oslo_vmware.api [None req-decfc0cf-abaa-40f8-a951-07480044abae tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Task: {'id': task-1117820, 'name': ResetVM_Task, 'duration_secs': 0.095974} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.491765] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-decfc0cf-abaa-40f8-a951-07480044abae tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Did hard reboot of VM {{(pid=63531) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1068}} [ 814.491932] env[63531]: DEBUG nova.compute.manager [None req-decfc0cf-abaa-40f8-a951-07480044abae tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 814.493596] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7d85671-e555-41c4-b786-878611ebae28 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.505652] env[63531]: DEBUG nova.compute.manager [req-fe564568-0338-4d76-8c36-6778826602eb req-e3b9abb6-9475-4aec-9843-6662a722f66c service nova] [instance: eb6c84d7-e0e6-4999-9018-889b99dc67ad] Received event network-changed-01bd4815-38ea-44f3-b7d3-c24fef3fd863 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 814.505919] env[63531]: DEBUG nova.compute.manager [req-fe564568-0338-4d76-8c36-6778826602eb req-e3b9abb6-9475-4aec-9843-6662a722f66c service nova] [instance: eb6c84d7-e0e6-4999-9018-889b99dc67ad] Refreshing instance network info cache due to event network-changed-01bd4815-38ea-44f3-b7d3-c24fef3fd863. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 814.506227] env[63531]: DEBUG oslo_concurrency.lockutils [req-fe564568-0338-4d76-8c36-6778826602eb req-e3b9abb6-9475-4aec-9843-6662a722f66c service nova] Acquiring lock "refresh_cache-eb6c84d7-e0e6-4999-9018-889b99dc67ad" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.506426] env[63531]: DEBUG oslo_concurrency.lockutils [req-fe564568-0338-4d76-8c36-6778826602eb req-e3b9abb6-9475-4aec-9843-6662a722f66c service nova] Acquired lock "refresh_cache-eb6c84d7-e0e6-4999-9018-889b99dc67ad" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.506669] env[63531]: DEBUG nova.network.neutron [req-fe564568-0338-4d76-8c36-6778826602eb req-e3b9abb6-9475-4aec-9843-6662a722f66c service nova] [instance: eb6c84d7-e0e6-4999-9018-889b99dc67ad] Refreshing network info cache for port 01bd4815-38ea-44f3-b7d3-c24fef3fd863 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 814.669341] env[63531]: DEBUG nova.objects.base [None req-bec51a9c-2a9e-4d84-8fff-a20a798f4290 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Object Instance<29dd6de3-2765-408c-acc0-da47e5e0a977> lazy-loaded attributes: flavor,info_cache {{(pid=63531) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 814.716164] env[63531]: DEBUG oslo_vmware.api [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1117819, 'name': PowerOnVM_Task, 'duration_secs': 0.710347} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.717801] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 531e1852-0c67-4b4d-a0ca-749ac438e79a] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 814.718154] env[63531]: INFO nova.compute.manager [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 531e1852-0c67-4b4d-a0ca-749ac438e79a] Took 10.06 seconds to spawn the instance on the hypervisor. [ 814.718424] env[63531]: DEBUG nova.compute.manager [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 531e1852-0c67-4b4d-a0ca-749ac438e79a] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 814.720606] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-370d56aa-16ea-4f17-919e-5a8df66ddac7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.750552] env[63531]: DEBUG nova.compute.manager [req-2cb9e5c9-eca0-4927-89e6-52e063fd3bf6 req-6601907f-c38e-4aca-9b94-032fbec3edac service nova] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Received event network-changed-2d27a8a5-c288-44fe-ac09-0d3b488386d6 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 814.750756] env[63531]: DEBUG nova.compute.manager [req-2cb9e5c9-eca0-4927-89e6-52e063fd3bf6 req-6601907f-c38e-4aca-9b94-032fbec3edac service nova] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Refreshing instance network info cache due to event network-changed-2d27a8a5-c288-44fe-ac09-0d3b488386d6. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 814.750986] env[63531]: DEBUG oslo_concurrency.lockutils [req-2cb9e5c9-eca0-4927-89e6-52e063fd3bf6 req-6601907f-c38e-4aca-9b94-032fbec3edac service nova] Acquiring lock "refresh_cache-2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.751160] env[63531]: DEBUG oslo_concurrency.lockutils [req-2cb9e5c9-eca0-4927-89e6-52e063fd3bf6 req-6601907f-c38e-4aca-9b94-032fbec3edac service nova] Acquired lock "refresh_cache-2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.751332] env[63531]: DEBUG nova.network.neutron [req-2cb9e5c9-eca0-4927-89e6-52e063fd3bf6 req-6601907f-c38e-4aca-9b94-032fbec3edac service nova] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Refreshing network info cache for port 2d27a8a5-c288-44fe-ac09-0d3b488386d6 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 814.782500] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117823, 'name': CreateVM_Task, 'duration_secs': 0.342682} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.782633] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eb6c84d7-e0e6-4999-9018-889b99dc67ad] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 814.783350] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.783517] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.783847] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 814.784187] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e724879a-f589-4263-8a20-25a377bb255b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.794199] env[63531]: DEBUG oslo_vmware.api [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Waiting for the task: (returnval){ [ 814.794199] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52545016-ec8d-eaf8-5f1b-8f81b63dc6e8" [ 814.794199] env[63531]: _type = "Task" [ 814.794199] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.802840] env[63531]: DEBUG oslo_vmware.api [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52545016-ec8d-eaf8-5f1b-8f81b63dc6e8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.018564] env[63531]: DEBUG oslo_concurrency.lockutils [None req-decfc0cf-abaa-40f8-a951-07480044abae tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Lock "caf0dee8-cb5a-402b-b533-22a590433f54" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.253s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.070275] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c574199-f63b-4075-902c-5ebdf28c3f2d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.080499] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b236e9d-3b95-473b-ace0-31a0cd657d32 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.115981] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-267d1409-434d-49ed-865a-7687df43c098 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.125855] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f082d1f-6fae-4e7f-b4cd-087d5dd30aad {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.144762] env[63531]: DEBUG nova.compute.provider_tree [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 815.245387] env[63531]: INFO nova.compute.manager [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 531e1852-0c67-4b4d-a0ca-749ac438e79a] Took 42.34 seconds to build instance. [ 815.305562] env[63531]: DEBUG oslo_vmware.api [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52545016-ec8d-eaf8-5f1b-8f81b63dc6e8, 'name': SearchDatastore_Task, 'duration_secs': 0.014024} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.305874] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.306127] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] [instance: eb6c84d7-e0e6-4999-9018-889b99dc67ad] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 815.306375] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 815.306548] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.306708] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 815.306979] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e9ffcd70-90c6-4652-bb20-41a76ae9b159 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.316221] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 815.316428] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 815.317452] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cafb71e8-493e-4f65-b966-09a889cf72d6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.324143] env[63531]: DEBUG oslo_vmware.api [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Waiting for the task: (returnval){ [ 815.324143] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52e43efd-3c66-1027-5f54-e1051e97223c" [ 815.324143] env[63531]: _type = "Task" [ 815.324143] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.335620] env[63531]: DEBUG oslo_vmware.api [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52e43efd-3c66-1027-5f54-e1051e97223c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.367249] env[63531]: DEBUG nova.network.neutron [req-fe564568-0338-4d76-8c36-6778826602eb req-e3b9abb6-9475-4aec-9843-6662a722f66c service nova] [instance: eb6c84d7-e0e6-4999-9018-889b99dc67ad] Updated VIF entry in instance network info cache for port 01bd4815-38ea-44f3-b7d3-c24fef3fd863. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 815.367735] env[63531]: DEBUG nova.network.neutron [req-fe564568-0338-4d76-8c36-6778826602eb req-e3b9abb6-9475-4aec-9843-6662a722f66c service nova] [instance: eb6c84d7-e0e6-4999-9018-889b99dc67ad] Updating instance_info_cache with network_info: [{"id": "01bd4815-38ea-44f3-b7d3-c24fef3fd863", "address": "fa:16:3e:50:3e:97", "network": {"id": "c818e149-d1f0-4573-a514-996cb03d2e96", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-372309638-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1093a53d17004ed0984d01aed7487993", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1e029825-6c65-4ac7-88f6-65f9d106db76", "external-id": "nsx-vlan-transportzone-428", "segmentation_id": 428, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01bd4815-38", "ovs_interfaceid": "01bd4815-38ea-44f3-b7d3-c24fef3fd863", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 815.531364] env[63531]: DEBUG nova.network.neutron [None req-bec51a9c-2a9e-4d84-8fff-a20a798f4290 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Updating instance_info_cache with network_info: [{"id": "262abfb7-7bcf-40e6-9ce3-b502de45b3b4", "address": "fa:16:3e:0e:00:c9", "network": {"id": "0667c176-cfdf-4e43-b152-e00bdc224448", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-243715985-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.217", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "308eabafea9c4188a58a0f1c22074d2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8f580e6-1d86-41ee-9ebe-c531cb9299c6", "external-id": "nsx-vlan-transportzone-150", "segmentation_id": 150, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap262abfb7-7b", "ovs_interfaceid": "262abfb7-7bcf-40e6-9ce3-b502de45b3b4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 815.631361] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquiring lock "4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 815.631668] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Lock "4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 815.651780] env[63531]: DEBUG nova.scheduler.client.report [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 815.753131] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9726f8ec-4356-41f3-a3ac-4455adc97a7e tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Lock "531e1852-0c67-4b4d-a0ca-749ac438e79a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 68.570s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.843917] env[63531]: DEBUG oslo_vmware.api [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52e43efd-3c66-1027-5f54-e1051e97223c, 'name': SearchDatastore_Task, 'duration_secs': 0.015644} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.844414] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-673800c2-55ac-4ff9-8678-9af85e28b257 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.854229] env[63531]: DEBUG oslo_vmware.api [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Waiting for the task: (returnval){ [ 815.854229] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52efc343-5ad0-4ff9-d4d3-d516dfbe28d0" [ 815.854229] env[63531]: _type = "Task" [ 815.854229] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.864806] env[63531]: DEBUG oslo_vmware.api [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52efc343-5ad0-4ff9-d4d3-d516dfbe28d0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.872616] env[63531]: DEBUG oslo_concurrency.lockutils [req-fe564568-0338-4d76-8c36-6778826602eb req-e3b9abb6-9475-4aec-9843-6662a722f66c service nova] Releasing lock "refresh_cache-eb6c84d7-e0e6-4999-9018-889b99dc67ad" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.995727] env[63531]: DEBUG nova.network.neutron [req-2cb9e5c9-eca0-4927-89e6-52e063fd3bf6 req-6601907f-c38e-4aca-9b94-032fbec3edac service nova] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Updated VIF entry in instance network info cache for port 2d27a8a5-c288-44fe-ac09-0d3b488386d6. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 815.996120] env[63531]: DEBUG nova.network.neutron [req-2cb9e5c9-eca0-4927-89e6-52e063fd3bf6 req-6601907f-c38e-4aca-9b94-032fbec3edac service nova] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Updating instance_info_cache with network_info: [{"id": "2d27a8a5-c288-44fe-ac09-0d3b488386d6", "address": "fa:16:3e:91:1c:62", "network": {"id": "5b8b7134-f123-40b8-9134-5cd072e8ad01", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1232565840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.165", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef3b56ce3f3a4155bd09642f5c48ea6f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2d27a8a5-c2", "ovs_interfaceid": "2d27a8a5-c288-44fe-ac09-0d3b488386d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.018333] env[63531]: DEBUG nova.network.neutron [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Successfully updated port: a3fd640b-c47b-4501-ac04-f90f6af8b076 {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 816.037203] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bec51a9c-2a9e-4d84-8fff-a20a798f4290 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Releasing lock "refresh_cache-29dd6de3-2765-408c-acc0-da47e5e0a977" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.158037] env[63531]: DEBUG oslo_concurrency.lockutils [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.568s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.158632] env[63531]: DEBUG nova.compute.manager [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 816.161580] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 27.925s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.255710] env[63531]: DEBUG nova.compute.manager [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 816.368960] env[63531]: DEBUG oslo_vmware.api [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52efc343-5ad0-4ff9-d4d3-d516dfbe28d0, 'name': SearchDatastore_Task, 'duration_secs': 0.01292} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.369358] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.369515] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] eb6c84d7-e0e6-4999-9018-889b99dc67ad/eb6c84d7-e0e6-4999-9018-889b99dc67ad.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 816.369787] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-94296d32-ae62-481e-9890-d4787a597e4e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.376404] env[63531]: DEBUG oslo_vmware.api [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Waiting for the task: (returnval){ [ 816.376404] env[63531]: value = "task-1117824" [ 816.376404] env[63531]: _type = "Task" [ 816.376404] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.384790] env[63531]: DEBUG oslo_vmware.api [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Task: {'id': task-1117824, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.499478] env[63531]: DEBUG oslo_concurrency.lockutils [req-2cb9e5c9-eca0-4927-89e6-52e063fd3bf6 req-6601907f-c38e-4aca-9b94-032fbec3edac service nova] Releasing lock "refresh_cache-2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.521589] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Acquiring lock "refresh_cache-f46d8a85-6ab7-436a-bc5e-40d0f80974e5" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.521946] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Acquired lock "refresh_cache-f46d8a85-6ab7-436a-bc5e-40d0f80974e5" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.522156] env[63531]: DEBUG nova.network.neutron [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 816.541169] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-bec51a9c-2a9e-4d84-8fff-a20a798f4290 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 816.541729] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e140a54d-8cdc-4a1f-899d-82496890ca38 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.549644] env[63531]: DEBUG oslo_vmware.api [None req-bec51a9c-2a9e-4d84-8fff-a20a798f4290 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 816.549644] env[63531]: value = "task-1117825" [ 816.549644] env[63531]: _type = "Task" [ 816.549644] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.558816] env[63531]: DEBUG oslo_vmware.api [None req-bec51a9c-2a9e-4d84-8fff-a20a798f4290 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1117825, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.625104] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquiring lock "879924d4-b465-4102-a0e3-c7b2be7ef08b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.625353] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Lock "879924d4-b465-4102-a0e3-c7b2be7ef08b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.665770] env[63531]: DEBUG nova.compute.utils [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 816.667696] env[63531]: DEBUG nova.objects.instance [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Lazy-loading 'migration_context' on Instance uuid 5e9042b3-4dc2-4fa3-a664-c4b49a22e400 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 816.669168] env[63531]: DEBUG nova.compute.manager [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 816.669754] env[63531]: DEBUG nova.network.neutron [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 816.735862] env[63531]: DEBUG nova.policy [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '69131938118a405b9bb362be787a3f14', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '824ed405cc6e4868a70e82489e05e8ac', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 816.777106] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.887776] env[63531]: DEBUG oslo_vmware.api [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Task: {'id': task-1117824, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.063194] env[63531]: DEBUG oslo_vmware.api [None req-bec51a9c-2a9e-4d84-8fff-a20a798f4290 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1117825, 'name': PowerOnVM_Task, 'duration_secs': 0.451236} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.063194] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-bec51a9c-2a9e-4d84-8fff-a20a798f4290 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 817.063194] env[63531]: DEBUG nova.compute.manager [None req-bec51a9c-2a9e-4d84-8fff-a20a798f4290 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 817.063194] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9317563-7fb2-4010-a867-2bca00f7598b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.096803] env[63531]: DEBUG nova.compute.manager [req-ec2ecf47-8bb6-4374-9855-1813dd0fec35 req-543c34ea-5b4b-40ff-8e45-de7f9e6c73ac service nova] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Received event network-vif-plugged-a3fd640b-c47b-4501-ac04-f90f6af8b076 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 817.096803] env[63531]: DEBUG oslo_concurrency.lockutils [req-ec2ecf47-8bb6-4374-9855-1813dd0fec35 req-543c34ea-5b4b-40ff-8e45-de7f9e6c73ac service nova] Acquiring lock "f46d8a85-6ab7-436a-bc5e-40d0f80974e5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.096803] env[63531]: DEBUG oslo_concurrency.lockutils [req-ec2ecf47-8bb6-4374-9855-1813dd0fec35 req-543c34ea-5b4b-40ff-8e45-de7f9e6c73ac service nova] Lock "f46d8a85-6ab7-436a-bc5e-40d0f80974e5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 817.097279] env[63531]: DEBUG oslo_concurrency.lockutils [req-ec2ecf47-8bb6-4374-9855-1813dd0fec35 req-543c34ea-5b4b-40ff-8e45-de7f9e6c73ac service nova] Lock "f46d8a85-6ab7-436a-bc5e-40d0f80974e5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.097670] env[63531]: DEBUG nova.compute.manager [req-ec2ecf47-8bb6-4374-9855-1813dd0fec35 req-543c34ea-5b4b-40ff-8e45-de7f9e6c73ac service nova] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] No waiting events found dispatching network-vif-plugged-a3fd640b-c47b-4501-ac04-f90f6af8b076 {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 817.098122] env[63531]: WARNING nova.compute.manager [req-ec2ecf47-8bb6-4374-9855-1813dd0fec35 req-543c34ea-5b4b-40ff-8e45-de7f9e6c73ac service nova] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Received unexpected event network-vif-plugged-a3fd640b-c47b-4501-ac04-f90f6af8b076 for instance with vm_state building and task_state spawning. [ 817.098428] env[63531]: DEBUG nova.compute.manager [req-ec2ecf47-8bb6-4374-9855-1813dd0fec35 req-543c34ea-5b4b-40ff-8e45-de7f9e6c73ac service nova] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Received event network-changed-a3fd640b-c47b-4501-ac04-f90f6af8b076 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 817.099076] env[63531]: DEBUG nova.compute.manager [req-ec2ecf47-8bb6-4374-9855-1813dd0fec35 req-543c34ea-5b4b-40ff-8e45-de7f9e6c73ac service nova] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Refreshing instance network info cache due to event network-changed-a3fd640b-c47b-4501-ac04-f90f6af8b076. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 817.099810] env[63531]: DEBUG oslo_concurrency.lockutils [req-ec2ecf47-8bb6-4374-9855-1813dd0fec35 req-543c34ea-5b4b-40ff-8e45-de7f9e6c73ac service nova] Acquiring lock "refresh_cache-f46d8a85-6ab7-436a-bc5e-40d0f80974e5" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.107268] env[63531]: DEBUG nova.network.neutron [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 817.172033] env[63531]: DEBUG nova.compute.manager [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 817.221393] env[63531]: DEBUG oslo_concurrency.lockutils [None req-34c5a484-e724-4b75-a347-4f418835aae3 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Acquiring lock "531e1852-0c67-4b4d-a0ca-749ac438e79a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.224255] env[63531]: DEBUG oslo_concurrency.lockutils [None req-34c5a484-e724-4b75-a347-4f418835aae3 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Lock "531e1852-0c67-4b4d-a0ca-749ac438e79a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 817.224255] env[63531]: DEBUG oslo_concurrency.lockutils [None req-34c5a484-e724-4b75-a347-4f418835aae3 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Acquiring lock "531e1852-0c67-4b4d-a0ca-749ac438e79a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.224255] env[63531]: DEBUG oslo_concurrency.lockutils [None req-34c5a484-e724-4b75-a347-4f418835aae3 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Lock "531e1852-0c67-4b4d-a0ca-749ac438e79a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 817.224255] env[63531]: DEBUG oslo_concurrency.lockutils [None req-34c5a484-e724-4b75-a347-4f418835aae3 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Lock "531e1852-0c67-4b4d-a0ca-749ac438e79a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.235777] env[63531]: INFO nova.compute.manager [None req-34c5a484-e724-4b75-a347-4f418835aae3 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 531e1852-0c67-4b4d-a0ca-749ac438e79a] Terminating instance [ 817.239168] env[63531]: DEBUG nova.compute.manager [None req-34c5a484-e724-4b75-a347-4f418835aae3 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 531e1852-0c67-4b4d-a0ca-749ac438e79a] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 817.239251] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-34c5a484-e724-4b75-a347-4f418835aae3 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 531e1852-0c67-4b4d-a0ca-749ac438e79a] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 817.240131] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fd16a3f-32dd-43d7-930f-4be09c2700c4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.248649] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-34c5a484-e724-4b75-a347-4f418835aae3 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 531e1852-0c67-4b4d-a0ca-749ac438e79a] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 817.248649] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-73ffd478-43de-4037-a1fc-f506a4dd4510 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.259284] env[63531]: DEBUG oslo_vmware.api [None req-34c5a484-e724-4b75-a347-4f418835aae3 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Waiting for the task: (returnval){ [ 817.259284] env[63531]: value = "task-1117826" [ 817.259284] env[63531]: _type = "Task" [ 817.259284] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.268837] env[63531]: DEBUG oslo_vmware.api [None req-34c5a484-e724-4b75-a347-4f418835aae3 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1117826, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.387434] env[63531]: DEBUG oslo_vmware.api [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Task: {'id': task-1117824, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.458895] env[63531]: DEBUG oslo_concurrency.lockutils [None req-056f71aa-40aa-436b-8ae3-9a3a30a67adc tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Acquiring lock "caf0dee8-cb5a-402b-b533-22a590433f54" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.459332] env[63531]: DEBUG oslo_concurrency.lockutils [None req-056f71aa-40aa-436b-8ae3-9a3a30a67adc tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Lock "caf0dee8-cb5a-402b-b533-22a590433f54" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 817.459661] env[63531]: DEBUG oslo_concurrency.lockutils [None req-056f71aa-40aa-436b-8ae3-9a3a30a67adc tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Acquiring lock "caf0dee8-cb5a-402b-b533-22a590433f54-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.459925] env[63531]: DEBUG oslo_concurrency.lockutils [None req-056f71aa-40aa-436b-8ae3-9a3a30a67adc tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Lock "caf0dee8-cb5a-402b-b533-22a590433f54-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 817.460217] env[63531]: DEBUG oslo_concurrency.lockutils [None req-056f71aa-40aa-436b-8ae3-9a3a30a67adc tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Lock "caf0dee8-cb5a-402b-b533-22a590433f54-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.462537] env[63531]: INFO nova.compute.manager [None req-056f71aa-40aa-436b-8ae3-9a3a30a67adc tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Terminating instance [ 817.468806] env[63531]: DEBUG nova.compute.manager [None req-056f71aa-40aa-436b-8ae3-9a3a30a67adc tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 817.468806] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-056f71aa-40aa-436b-8ae3-9a3a30a67adc tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 817.468806] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58504069-9c76-4dbe-bd9e-92f00e088ac3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.477037] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-056f71aa-40aa-436b-8ae3-9a3a30a67adc tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 817.480220] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8cb63390-f08e-4390-9094-061dccb633cc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.491089] env[63531]: DEBUG oslo_vmware.api [None req-056f71aa-40aa-436b-8ae3-9a3a30a67adc tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Waiting for the task: (returnval){ [ 817.491089] env[63531]: value = "task-1117827" [ 817.491089] env[63531]: _type = "Task" [ 817.491089] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.492176] env[63531]: DEBUG nova.network.neutron [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Successfully created port: f1ef6f0e-1f62-4034-8edb-e8cc1bb98abd {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 817.509311] env[63531]: DEBUG nova.network.neutron [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Updating instance_info_cache with network_info: [{"id": "a3fd640b-c47b-4501-ac04-f90f6af8b076", "address": "fa:16:3e:27:ca:51", "network": {"id": "4dd26082-3411-4d99-a78b-c76c15703506", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1190890625-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ee5fd84e1da04fd984f81578e4085850", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6046aec4-feda-4ef9-bf4a-800de1e0cd3b", "external-id": "nsx-vlan-transportzone-903", "segmentation_id": 903, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3fd640b-c4", "ovs_interfaceid": "a3fd640b-c47b-4501-ac04-f90f6af8b076", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.513965] env[63531]: DEBUG oslo_vmware.api [None req-056f71aa-40aa-436b-8ae3-9a3a30a67adc tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Task: {'id': task-1117827, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.718014] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4afb5492-807a-4065-87db-7c5087ad00c4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.730919] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0870a4e-ed81-4c30-a4c5-993208c28edb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.767576] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d388ec5-be0c-4d16-b8c6-f20311cd35ac {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.774774] env[63531]: DEBUG oslo_vmware.api [None req-34c5a484-e724-4b75-a347-4f418835aae3 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1117826, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.777967] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ff92109-39b7-43af-96de-23ef1dc90053 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.792754] env[63531]: DEBUG nova.compute.provider_tree [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 817.888204] env[63531]: DEBUG oslo_vmware.api [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Task: {'id': task-1117824, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.016787] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Releasing lock "refresh_cache-f46d8a85-6ab7-436a-bc5e-40d0f80974e5" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.017257] env[63531]: DEBUG nova.compute.manager [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Instance network_info: |[{"id": "a3fd640b-c47b-4501-ac04-f90f6af8b076", "address": "fa:16:3e:27:ca:51", "network": {"id": "4dd26082-3411-4d99-a78b-c76c15703506", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1190890625-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ee5fd84e1da04fd984f81578e4085850", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6046aec4-feda-4ef9-bf4a-800de1e0cd3b", "external-id": "nsx-vlan-transportzone-903", "segmentation_id": 903, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3fd640b-c4", "ovs_interfaceid": "a3fd640b-c47b-4501-ac04-f90f6af8b076", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 818.017599] env[63531]: DEBUG oslo_vmware.api [None req-056f71aa-40aa-436b-8ae3-9a3a30a67adc tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Task: {'id': task-1117827, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.017861] env[63531]: DEBUG oslo_concurrency.lockutils [req-ec2ecf47-8bb6-4374-9855-1813dd0fec35 req-543c34ea-5b4b-40ff-8e45-de7f9e6c73ac service nova] Acquired lock "refresh_cache-f46d8a85-6ab7-436a-bc5e-40d0f80974e5" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.018098] env[63531]: DEBUG nova.network.neutron [req-ec2ecf47-8bb6-4374-9855-1813dd0fec35 req-543c34ea-5b4b-40ff-8e45-de7f9e6c73ac service nova] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Refreshing network info cache for port a3fd640b-c47b-4501-ac04-f90f6af8b076 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 818.020706] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:27:ca:51', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6046aec4-feda-4ef9-bf4a-800de1e0cd3b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a3fd640b-c47b-4501-ac04-f90f6af8b076', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 818.031941] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Creating folder: Project (ee5fd84e1da04fd984f81578e4085850). Parent ref: group-v244585. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 818.035188] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-695bb960-54cd-41b5-a0ef-ce6eb739a37c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.046080] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Created folder: Project (ee5fd84e1da04fd984f81578e4085850) in parent group-v244585. [ 818.046284] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Creating folder: Instances. Parent ref: group-v244684. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 818.046522] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3cf3b773-d7d3-4d7c-a6fc-b052273c088d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.055234] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Created folder: Instances in parent group-v244684. [ 818.055590] env[63531]: DEBUG oslo.service.loopingcall [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 818.055786] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 818.056376] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7695d0bb-5d11-41e8-a527-048ce8ed117c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.075707] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 818.075707] env[63531]: value = "task-1117830" [ 818.075707] env[63531]: _type = "Task" [ 818.075707] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.083724] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117830, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.190873] env[63531]: DEBUG nova.compute.manager [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 818.228017] env[63531]: DEBUG nova.virt.hardware [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 818.228346] env[63531]: DEBUG nova.virt.hardware [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 818.228520] env[63531]: DEBUG nova.virt.hardware [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 818.228711] env[63531]: DEBUG nova.virt.hardware [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 818.228866] env[63531]: DEBUG nova.virt.hardware [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 818.234446] env[63531]: DEBUG nova.virt.hardware [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 818.234891] env[63531]: DEBUG nova.virt.hardware [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 818.235150] env[63531]: DEBUG nova.virt.hardware [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 818.235365] env[63531]: DEBUG nova.virt.hardware [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 818.235546] env[63531]: DEBUG nova.virt.hardware [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 818.235749] env[63531]: DEBUG nova.virt.hardware [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 818.236719] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aa077d5-c196-42ca-91f5-c358e3aa5387 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.246355] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5e6aaf5-3d39-4d3e-8871-8e1f5dd51ad2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.279222] env[63531]: DEBUG oslo_vmware.api [None req-34c5a484-e724-4b75-a347-4f418835aae3 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1117826, 'name': PowerOffVM_Task} progress is 100%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.296447] env[63531]: DEBUG nova.scheduler.client.report [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 818.388933] env[63531]: DEBUG oslo_vmware.api [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Task: {'id': task-1117824, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.570227} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.389227] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] eb6c84d7-e0e6-4999-9018-889b99dc67ad/eb6c84d7-e0e6-4999-9018-889b99dc67ad.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 818.389493] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] [instance: eb6c84d7-e0e6-4999-9018-889b99dc67ad] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 818.389779] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3aba795a-07b6-4d37-a8d4-49a8165bee99 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.398582] env[63531]: DEBUG oslo_vmware.api [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Waiting for the task: (returnval){ [ 818.398582] env[63531]: value = "task-1117831" [ 818.398582] env[63531]: _type = "Task" [ 818.398582] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.405966] env[63531]: DEBUG oslo_vmware.api [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Task: {'id': task-1117831, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.510279] env[63531]: DEBUG oslo_vmware.api [None req-056f71aa-40aa-436b-8ae3-9a3a30a67adc tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Task: {'id': task-1117827, 'name': PowerOffVM_Task, 'duration_secs': 0.815548} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.510578] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-056f71aa-40aa-436b-8ae3-9a3a30a67adc tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 818.510750] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-056f71aa-40aa-436b-8ae3-9a3a30a67adc tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 818.511022] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5cdf9ebe-d434-4780-9032-74799dd2be50 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.587280] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117830, 'name': CreateVM_Task, 'duration_secs': 0.400172} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.587576] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 818.588242] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 818.588451] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.588797] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 818.589078] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-50e2bd8a-032b-4205-bc91-7e81eab3b0e6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.594318] env[63531]: DEBUG oslo_vmware.api [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Waiting for the task: (returnval){ [ 818.594318] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52d3de04-4a6d-f6fe-c4bf-33bc7312dfd5" [ 818.594318] env[63531]: _type = "Task" [ 818.594318] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.603713] env[63531]: DEBUG oslo_vmware.api [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52d3de04-4a6d-f6fe-c4bf-33bc7312dfd5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.618780] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-056f71aa-40aa-436b-8ae3-9a3a30a67adc tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 818.619048] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-056f71aa-40aa-436b-8ae3-9a3a30a67adc tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 818.619331] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-056f71aa-40aa-436b-8ae3-9a3a30a67adc tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Deleting the datastore file [datastore1] caf0dee8-cb5a-402b-b533-22a590433f54 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 818.619547] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-35841c8f-d844-44b3-9311-a0e42f27dc66 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.626510] env[63531]: DEBUG oslo_vmware.api [None req-056f71aa-40aa-436b-8ae3-9a3a30a67adc tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Waiting for the task: (returnval){ [ 818.626510] env[63531]: value = "task-1117833" [ 818.626510] env[63531]: _type = "Task" [ 818.626510] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.634555] env[63531]: DEBUG oslo_vmware.api [None req-056f71aa-40aa-436b-8ae3-9a3a30a67adc tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Task: {'id': task-1117833, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.687096] env[63531]: DEBUG nova.network.neutron [req-ec2ecf47-8bb6-4374-9855-1813dd0fec35 req-543c34ea-5b4b-40ff-8e45-de7f9e6c73ac service nova] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Updated VIF entry in instance network info cache for port a3fd640b-c47b-4501-ac04-f90f6af8b076. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 818.687563] env[63531]: DEBUG nova.network.neutron [req-ec2ecf47-8bb6-4374-9855-1813dd0fec35 req-543c34ea-5b4b-40ff-8e45-de7f9e6c73ac service nova] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Updating instance_info_cache with network_info: [{"id": "a3fd640b-c47b-4501-ac04-f90f6af8b076", "address": "fa:16:3e:27:ca:51", "network": {"id": "4dd26082-3411-4d99-a78b-c76c15703506", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1190890625-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ee5fd84e1da04fd984f81578e4085850", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6046aec4-feda-4ef9-bf4a-800de1e0cd3b", "external-id": "nsx-vlan-transportzone-903", "segmentation_id": 903, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3fd640b-c4", "ovs_interfaceid": "a3fd640b-c47b-4501-ac04-f90f6af8b076", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 818.776652] env[63531]: DEBUG oslo_vmware.api [None req-34c5a484-e724-4b75-a347-4f418835aae3 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1117826, 'name': PowerOffVM_Task, 'duration_secs': 1.044173} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.776982] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-34c5a484-e724-4b75-a347-4f418835aae3 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 531e1852-0c67-4b4d-a0ca-749ac438e79a] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 818.777294] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-34c5a484-e724-4b75-a347-4f418835aae3 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 531e1852-0c67-4b4d-a0ca-749ac438e79a] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 818.777583] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4f6039f7-aa2d-429e-b10a-24af8397db9b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.850066] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-34c5a484-e724-4b75-a347-4f418835aae3 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 531e1852-0c67-4b4d-a0ca-749ac438e79a] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 818.850066] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-34c5a484-e724-4b75-a347-4f418835aae3 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 531e1852-0c67-4b4d-a0ca-749ac438e79a] Deleting contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 818.850893] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-34c5a484-e724-4b75-a347-4f418835aae3 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Deleting the datastore file [datastore2] 531e1852-0c67-4b4d-a0ca-749ac438e79a {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 818.850893] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-604437cd-8059-4bac-899d-229f98574c5b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.857289] env[63531]: DEBUG oslo_vmware.api [None req-34c5a484-e724-4b75-a347-4f418835aae3 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Waiting for the task: (returnval){ [ 818.857289] env[63531]: value = "task-1117835" [ 818.857289] env[63531]: _type = "Task" [ 818.857289] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.867141] env[63531]: DEBUG oslo_vmware.api [None req-34c5a484-e724-4b75-a347-4f418835aae3 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1117835, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.908605] env[63531]: DEBUG oslo_vmware.api [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Task: {'id': task-1117831, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072651} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.908925] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] [instance: eb6c84d7-e0e6-4999-9018-889b99dc67ad] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 818.909786] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f736acd-a22a-40ee-ad0c-4d25885f72ad {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.933923] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] [instance: eb6c84d7-e0e6-4999-9018-889b99dc67ad] Reconfiguring VM instance instance-0000003b to attach disk [datastore2] eb6c84d7-e0e6-4999-9018-889b99dc67ad/eb6c84d7-e0e6-4999-9018-889b99dc67ad.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 818.934371] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6086fcab-a823-4e4a-ac95-29f1cdf7f191 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.956216] env[63531]: DEBUG oslo_vmware.api [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Waiting for the task: (returnval){ [ 818.956216] env[63531]: value = "task-1117836" [ 818.956216] env[63531]: _type = "Task" [ 818.956216] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.966903] env[63531]: DEBUG oslo_vmware.api [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Task: {'id': task-1117836, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.104659] env[63531]: DEBUG oslo_vmware.api [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52d3de04-4a6d-f6fe-c4bf-33bc7312dfd5, 'name': SearchDatastore_Task, 'duration_secs': 0.016561} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.104969] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.105236] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 819.105536] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.105696] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.105885] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 819.106162] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5f52f473-a24f-43b7-8a35-0905e3d06652 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.114176] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 819.114388] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 819.115061] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f720416d-13f7-4e00-9861-ec9d0f6ba0f4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.120440] env[63531]: DEBUG oslo_vmware.api [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Waiting for the task: (returnval){ [ 819.120440] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52324e47-a327-6c72-303e-6e45fa8aa44c" [ 819.120440] env[63531]: _type = "Task" [ 819.120440] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.132970] env[63531]: DEBUG oslo_vmware.api [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52324e47-a327-6c72-303e-6e45fa8aa44c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.139465] env[63531]: DEBUG oslo_vmware.api [None req-056f71aa-40aa-436b-8ae3-9a3a30a67adc tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Task: {'id': task-1117833, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.245004} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.139562] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-056f71aa-40aa-436b-8ae3-9a3a30a67adc tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 819.139796] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-056f71aa-40aa-436b-8ae3-9a3a30a67adc tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 819.140118] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-056f71aa-40aa-436b-8ae3-9a3a30a67adc tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 819.140435] env[63531]: INFO nova.compute.manager [None req-056f71aa-40aa-436b-8ae3-9a3a30a67adc tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Took 1.67 seconds to destroy the instance on the hypervisor. [ 819.140804] env[63531]: DEBUG oslo.service.loopingcall [None req-056f71aa-40aa-436b-8ae3-9a3a30a67adc tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 819.141124] env[63531]: DEBUG nova.compute.manager [-] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 819.141304] env[63531]: DEBUG nova.network.neutron [-] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 819.193161] env[63531]: DEBUG oslo_concurrency.lockutils [req-ec2ecf47-8bb6-4374-9855-1813dd0fec35 req-543c34ea-5b4b-40ff-8e45-de7f9e6c73ac service nova] Releasing lock "refresh_cache-f46d8a85-6ab7-436a-bc5e-40d0f80974e5" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.193161] env[63531]: DEBUG nova.compute.manager [req-ec2ecf47-8bb6-4374-9855-1813dd0fec35 req-543c34ea-5b4b-40ff-8e45-de7f9e6c73ac service nova] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Received event network-changed-d99cda7f-476b-4869-8ba7-4a00c03da394 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 819.193161] env[63531]: DEBUG nova.compute.manager [req-ec2ecf47-8bb6-4374-9855-1813dd0fec35 req-543c34ea-5b4b-40ff-8e45-de7f9e6c73ac service nova] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Refreshing instance network info cache due to event network-changed-d99cda7f-476b-4869-8ba7-4a00c03da394. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 819.193161] env[63531]: DEBUG oslo_concurrency.lockutils [req-ec2ecf47-8bb6-4374-9855-1813dd0fec35 req-543c34ea-5b4b-40ff-8e45-de7f9e6c73ac service nova] Acquiring lock "refresh_cache-caf0dee8-cb5a-402b-b533-22a590433f54" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.193161] env[63531]: DEBUG oslo_concurrency.lockutils [req-ec2ecf47-8bb6-4374-9855-1813dd0fec35 req-543c34ea-5b4b-40ff-8e45-de7f9e6c73ac service nova] Acquired lock "refresh_cache-caf0dee8-cb5a-402b-b533-22a590433f54" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.193161] env[63531]: DEBUG nova.network.neutron [req-ec2ecf47-8bb6-4374-9855-1813dd0fec35 req-543c34ea-5b4b-40ff-8e45-de7f9e6c73ac service nova] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Refreshing network info cache for port d99cda7f-476b-4869-8ba7-4a00c03da394 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 819.315638] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 3.154s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.329084] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9e77b61d-e87e-4223-8227-0c685c107554 tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.015s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.329367] env[63531]: DEBUG nova.objects.instance [None req-9e77b61d-e87e-4223-8227-0c685c107554 tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Lazy-loading 'resources' on Instance uuid 410ddf7a-e45b-4769-a3db-5363ce2096a7 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 819.369082] env[63531]: DEBUG oslo_vmware.api [None req-34c5a484-e724-4b75-a347-4f418835aae3 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1117835, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.242386} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.369082] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-34c5a484-e724-4b75-a347-4f418835aae3 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 819.369270] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-34c5a484-e724-4b75-a347-4f418835aae3 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 531e1852-0c67-4b4d-a0ca-749ac438e79a] Deleted contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 819.369315] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-34c5a484-e724-4b75-a347-4f418835aae3 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 531e1852-0c67-4b4d-a0ca-749ac438e79a] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 819.369468] env[63531]: INFO nova.compute.manager [None req-34c5a484-e724-4b75-a347-4f418835aae3 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 531e1852-0c67-4b4d-a0ca-749ac438e79a] Took 2.13 seconds to destroy the instance on the hypervisor. [ 819.369711] env[63531]: DEBUG oslo.service.loopingcall [None req-34c5a484-e724-4b75-a347-4f418835aae3 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 819.369895] env[63531]: DEBUG nova.compute.manager [-] [instance: 531e1852-0c67-4b4d-a0ca-749ac438e79a] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 819.370017] env[63531]: DEBUG nova.network.neutron [-] [instance: 531e1852-0c67-4b4d-a0ca-749ac438e79a] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 819.469585] env[63531]: DEBUG oslo_vmware.api [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Task: {'id': task-1117836, 'name': ReconfigVM_Task, 'duration_secs': 0.441582} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.469585] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] [instance: eb6c84d7-e0e6-4999-9018-889b99dc67ad] Reconfigured VM instance instance-0000003b to attach disk [datastore2] eb6c84d7-e0e6-4999-9018-889b99dc67ad/eb6c84d7-e0e6-4999-9018-889b99dc67ad.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 819.470202] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c85d0a65-b69b-4771-bbc2-96c7c2408b0e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.476958] env[63531]: DEBUG oslo_vmware.api [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Waiting for the task: (returnval){ [ 819.476958] env[63531]: value = "task-1117837" [ 819.476958] env[63531]: _type = "Task" [ 819.476958] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.487715] env[63531]: DEBUG oslo_vmware.api [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Task: {'id': task-1117837, 'name': Rename_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.633792] env[63531]: DEBUG oslo_vmware.api [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52324e47-a327-6c72-303e-6e45fa8aa44c, 'name': SearchDatastore_Task, 'duration_secs': 0.011388} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.634724] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f6d363ec-3d50-4373-b092-07b7be23b58e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.641037] env[63531]: DEBUG oslo_vmware.api [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Waiting for the task: (returnval){ [ 819.641037] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]524d64ce-26ba-aec5-dd6c-03b331ef42e7" [ 819.641037] env[63531]: _type = "Task" [ 819.641037] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.661271] env[63531]: DEBUG oslo_vmware.api [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]524d64ce-26ba-aec5-dd6c-03b331ef42e7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.769648] env[63531]: DEBUG oslo_concurrency.lockutils [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquiring lock "e95d3273-8216-47cc-95b6-99301366a827" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.769772] env[63531]: DEBUG oslo_concurrency.lockutils [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "e95d3273-8216-47cc-95b6-99301366a827" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.970517] env[63531]: DEBUG nova.compute.manager [req-56036d3d-77c2-4320-96da-88510eef893f req-75a1c70c-71bb-41d8-b9b9-06484be6ab25 service nova] [instance: 531e1852-0c67-4b4d-a0ca-749ac438e79a] Received event network-vif-deleted-99ca461b-afde-4ea5-8cbc-e596f9f8e528 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 819.970988] env[63531]: INFO nova.compute.manager [req-56036d3d-77c2-4320-96da-88510eef893f req-75a1c70c-71bb-41d8-b9b9-06484be6ab25 service nova] [instance: 531e1852-0c67-4b4d-a0ca-749ac438e79a] Neutron deleted interface 99ca461b-afde-4ea5-8cbc-e596f9f8e528; detaching it from the instance and deleting it from the info cache [ 819.971605] env[63531]: DEBUG nova.network.neutron [req-56036d3d-77c2-4320-96da-88510eef893f req-75a1c70c-71bb-41d8-b9b9-06484be6ab25 service nova] [instance: 531e1852-0c67-4b4d-a0ca-749ac438e79a] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 819.978311] env[63531]: DEBUG nova.network.neutron [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Successfully updated port: f1ef6f0e-1f62-4034-8edb-e8cc1bb98abd {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 819.989064] env[63531]: DEBUG nova.network.neutron [req-ec2ecf47-8bb6-4374-9855-1813dd0fec35 req-543c34ea-5b4b-40ff-8e45-de7f9e6c73ac service nova] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Updated VIF entry in instance network info cache for port d99cda7f-476b-4869-8ba7-4a00c03da394. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 819.989064] env[63531]: DEBUG nova.network.neutron [req-ec2ecf47-8bb6-4374-9855-1813dd0fec35 req-543c34ea-5b4b-40ff-8e45-de7f9e6c73ac service nova] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Updating instance_info_cache with network_info: [{"id": "d99cda7f-476b-4869-8ba7-4a00c03da394", "address": "fa:16:3e:9d:85:ab", "network": {"id": "067f0236-aa59-4a63-a33c-455dc28e0c1a", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-54491452-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a0e384d6cf184a98aa3602eca6b3f410", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7bdf594e-da7a-4254-b413-87aef4614588", "external-id": "nsx-vlan-transportzone-422", "segmentation_id": 422, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd99cda7f-47", "ovs_interfaceid": "d99cda7f-476b-4869-8ba7-4a00c03da394", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 819.994155] env[63531]: DEBUG oslo_vmware.api [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Task: {'id': task-1117837, 'name': Rename_Task, 'duration_secs': 0.15901} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.997778] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] [instance: eb6c84d7-e0e6-4999-9018-889b99dc67ad] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 820.000259] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-916da117-96a5-4bd6-b342-4e6adb4c404d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.008445] env[63531]: DEBUG oslo_vmware.api [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Waiting for the task: (returnval){ [ 820.008445] env[63531]: value = "task-1117838" [ 820.008445] env[63531]: _type = "Task" [ 820.008445] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.019901] env[63531]: DEBUG oslo_vmware.api [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Task: {'id': task-1117838, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.035729] env[63531]: DEBUG nova.compute.manager [req-81f7b1b7-a8e7-4847-a578-4c982b18c99e req-632c14a3-daea-409d-9c6b-16b44470ed69 service nova] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Received event network-vif-plugged-f1ef6f0e-1f62-4034-8edb-e8cc1bb98abd {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 820.036117] env[63531]: DEBUG oslo_concurrency.lockutils [req-81f7b1b7-a8e7-4847-a578-4c982b18c99e req-632c14a3-daea-409d-9c6b-16b44470ed69 service nova] Acquiring lock "e0074e2e-e8bd-4bcb-9ad9-271a10e079e4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.036274] env[63531]: DEBUG oslo_concurrency.lockutils [req-81f7b1b7-a8e7-4847-a578-4c982b18c99e req-632c14a3-daea-409d-9c6b-16b44470ed69 service nova] Lock "e0074e2e-e8bd-4bcb-9ad9-271a10e079e4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.036500] env[63531]: DEBUG oslo_concurrency.lockutils [req-81f7b1b7-a8e7-4847-a578-4c982b18c99e req-632c14a3-daea-409d-9c6b-16b44470ed69 service nova] Lock "e0074e2e-e8bd-4bcb-9ad9-271a10e079e4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.036696] env[63531]: DEBUG nova.compute.manager [req-81f7b1b7-a8e7-4847-a578-4c982b18c99e req-632c14a3-daea-409d-9c6b-16b44470ed69 service nova] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] No waiting events found dispatching network-vif-plugged-f1ef6f0e-1f62-4034-8edb-e8cc1bb98abd {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 820.036910] env[63531]: WARNING nova.compute.manager [req-81f7b1b7-a8e7-4847-a578-4c982b18c99e req-632c14a3-daea-409d-9c6b-16b44470ed69 service nova] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Received unexpected event network-vif-plugged-f1ef6f0e-1f62-4034-8edb-e8cc1bb98abd for instance with vm_state building and task_state spawning. [ 820.157243] env[63531]: DEBUG oslo_vmware.api [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]524d64ce-26ba-aec5-dd6c-03b331ef42e7, 'name': SearchDatastore_Task, 'duration_secs': 0.010327} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.157679] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.158136] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] f46d8a85-6ab7-436a-bc5e-40d0f80974e5/f46d8a85-6ab7-436a-bc5e-40d0f80974e5.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 820.158524] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-65b51b13-82b0-4358-8c67-9a127b5c9429 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.161696] env[63531]: DEBUG nova.network.neutron [-] [instance: 531e1852-0c67-4b4d-a0ca-749ac438e79a] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 820.171654] env[63531]: DEBUG oslo_vmware.api [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Waiting for the task: (returnval){ [ 820.171654] env[63531]: value = "task-1117839" [ 820.171654] env[63531]: _type = "Task" [ 820.171654] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.189653] env[63531]: DEBUG oslo_vmware.api [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1117839, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.322006] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6642826-70a9-4116-959f-4992adebcf28 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.330245] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55166997-077e-42b8-86db-f5ec660c699d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.369576] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04cbd3db-a462-4901-b154-fe0a2aff175f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.372737] env[63531]: DEBUG nova.network.neutron [-] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 820.379239] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22c80806-4126-4d62-b182-c11a03cbd381 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.395532] env[63531]: DEBUG nova.compute.provider_tree [None req-9e77b61d-e87e-4223-8227-0c685c107554 tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 820.477417] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b476551e-b610-4ba1-b30f-294f034657be {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.487422] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c752e801-5cbc-4f8b-a052-fa9d5c03fb6e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.498370] env[63531]: DEBUG oslo_concurrency.lockutils [req-ec2ecf47-8bb6-4374-9855-1813dd0fec35 req-543c34ea-5b4b-40ff-8e45-de7f9e6c73ac service nova] Releasing lock "refresh_cache-caf0dee8-cb5a-402b-b533-22a590433f54" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.499037] env[63531]: DEBUG oslo_concurrency.lockutils [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Acquiring lock "refresh_cache-e0074e2e-e8bd-4bcb-9ad9-271a10e079e4" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.499162] env[63531]: DEBUG oslo_concurrency.lockutils [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Acquired lock "refresh_cache-e0074e2e-e8bd-4bcb-9ad9-271a10e079e4" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.499309] env[63531]: DEBUG nova.network.neutron [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 820.525099] env[63531]: DEBUG nova.compute.manager [req-56036d3d-77c2-4320-96da-88510eef893f req-75a1c70c-71bb-41d8-b9b9-06484be6ab25 service nova] [instance: 531e1852-0c67-4b4d-a0ca-749ac438e79a] Detach interface failed, port_id=99ca461b-afde-4ea5-8cbc-e596f9f8e528, reason: Instance 531e1852-0c67-4b4d-a0ca-749ac438e79a could not be found. {{(pid=63531) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 820.532142] env[63531]: DEBUG oslo_vmware.api [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Task: {'id': task-1117838, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.668590] env[63531]: INFO nova.compute.manager [-] [instance: 531e1852-0c67-4b4d-a0ca-749ac438e79a] Took 1.30 seconds to deallocate network for instance. [ 820.688668] env[63531]: DEBUG oslo_vmware.api [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1117839, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.878068] env[63531]: INFO nova.compute.manager [-] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Took 1.74 seconds to deallocate network for instance. [ 820.895958] env[63531]: INFO nova.compute.manager [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Swapping old allocation on dict_keys(['9996a314-a62e-4282-8252-f387e7e3fabb']) held by migration 97923ce2-3d93-4ee9-a54b-c15bc8e3c7e3 for instance [ 820.900225] env[63531]: DEBUG nova.scheduler.client.report [None req-9e77b61d-e87e-4223-8227-0c685c107554 tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 820.928186] env[63531]: DEBUG nova.scheduler.client.report [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Overwriting current allocation {'allocations': {'9996a314-a62e-4282-8252-f387e7e3fabb': {'resources': {'VCPU': 1, 'MEMORY_MB': 256, 'DISK_GB': 1}, 'generation': 86}}, 'project_id': '5f0d10f679de4fa4bd9d725951e536c1', 'user_id': '2e047a775d4642dca798c5d624ccbc89', 'consumer_generation': 1} on consumer 5e9042b3-4dc2-4fa3-a664-c4b49a22e400 {{(pid=63531) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 821.019042] env[63531]: DEBUG oslo_vmware.api [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Task: {'id': task-1117838, 'name': PowerOnVM_Task, 'duration_secs': 0.703267} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.019521] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] [instance: eb6c84d7-e0e6-4999-9018-889b99dc67ad] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 821.019588] env[63531]: INFO nova.compute.manager [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] [instance: eb6c84d7-e0e6-4999-9018-889b99dc67ad] Took 9.65 seconds to spawn the instance on the hypervisor. [ 821.019723] env[63531]: DEBUG nova.compute.manager [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] [instance: eb6c84d7-e0e6-4999-9018-889b99dc67ad] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 821.020567] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8750f2fd-3354-4cac-8cd8-38e403d142f9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.023922] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Acquiring lock "refresh_cache-5e9042b3-4dc2-4fa3-a664-c4b49a22e400" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.024117] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Acquired lock "refresh_cache-5e9042b3-4dc2-4fa3-a664-c4b49a22e400" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.024298] env[63531]: DEBUG nova.network.neutron [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 821.046097] env[63531]: DEBUG nova.network.neutron [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 821.183534] env[63531]: DEBUG oslo_concurrency.lockutils [None req-34c5a484-e724-4b75-a347-4f418835aae3 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.190420] env[63531]: DEBUG oslo_vmware.api [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1117839, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.821672} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.190717] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] f46d8a85-6ab7-436a-bc5e-40d0f80974e5/f46d8a85-6ab7-436a-bc5e-40d0f80974e5.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 821.190970] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 821.191255] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-59343529-8450-46a6-b21e-d9de791eebb1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.198497] env[63531]: DEBUG oslo_vmware.api [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Waiting for the task: (returnval){ [ 821.198497] env[63531]: value = "task-1117840" [ 821.198497] env[63531]: _type = "Task" [ 821.198497] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.206258] env[63531]: DEBUG oslo_vmware.api [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1117840, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.283860] env[63531]: DEBUG nova.network.neutron [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Updating instance_info_cache with network_info: [{"id": "f1ef6f0e-1f62-4034-8edb-e8cc1bb98abd", "address": "fa:16:3e:3e:82:e5", "network": {"id": "dc416058-1c04-47ac-8b73-1557f7c33d1c", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1350689703-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "824ed405cc6e4868a70e82489e05e8ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "05b1253d-2b87-4158-9ff1-dafcf829f11f", "external-id": "nsx-vlan-transportzone-55", "segmentation_id": 55, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf1ef6f0e-1f", "ovs_interfaceid": "f1ef6f0e-1f62-4034-8edb-e8cc1bb98abd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.385311] env[63531]: DEBUG oslo_concurrency.lockutils [None req-056f71aa-40aa-436b-8ae3-9a3a30a67adc tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.404821] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9e77b61d-e87e-4223-8227-0c685c107554 tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.075s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.406727] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 32.787s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.408307] env[63531]: INFO nova.compute.claims [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] [instance: 43c29443-1210-4ee2-95c7-1257de308287] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 821.543135] env[63531]: INFO nova.compute.manager [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] [instance: eb6c84d7-e0e6-4999-9018-889b99dc67ad] Took 41.25 seconds to build instance. [ 821.708984] env[63531]: DEBUG oslo_vmware.api [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1117840, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.383776} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.711661] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 821.712575] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cd257fd-fe0b-4ffd-a757-50d827127c5f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.734810] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Reconfiguring VM instance instance-0000003c to attach disk [datastore2] f46d8a85-6ab7-436a-bc5e-40d0f80974e5/f46d8a85-6ab7-436a-bc5e-40d0f80974e5.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 821.735176] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-84a1e7b4-f93d-41f1-b74b-21756d0ddbcb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.755659] env[63531]: DEBUG oslo_vmware.api [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Waiting for the task: (returnval){ [ 821.755659] env[63531]: value = "task-1117841" [ 821.755659] env[63531]: _type = "Task" [ 821.755659] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.764149] env[63531]: DEBUG oslo_vmware.api [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1117841, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.765144] env[63531]: DEBUG nova.network.neutron [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Updating instance_info_cache with network_info: [{"id": "f178c862-e753-40d4-b7cd-3eabe7eda39e", "address": "fa:16:3e:a6:d3:d2", "network": {"id": "5439232a-ec6d-4184-a9f1-f7be5c1526bc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.35", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "54fbef8a6ee0488f820706d2036e9c99", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f267bcdd-0daa-4337-9709-5fc060c267d8", "external-id": "nsx-vlan-transportzone-308", "segmentation_id": 308, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf178c862-e7", "ovs_interfaceid": "f178c862-e753-40d4-b7cd-3eabe7eda39e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.786310] env[63531]: DEBUG oslo_concurrency.lockutils [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Releasing lock "refresh_cache-e0074e2e-e8bd-4bcb-9ad9-271a10e079e4" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.786923] env[63531]: DEBUG nova.compute.manager [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Instance network_info: |[{"id": "f1ef6f0e-1f62-4034-8edb-e8cc1bb98abd", "address": "fa:16:3e:3e:82:e5", "network": {"id": "dc416058-1c04-47ac-8b73-1557f7c33d1c", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1350689703-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "824ed405cc6e4868a70e82489e05e8ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "05b1253d-2b87-4158-9ff1-dafcf829f11f", "external-id": "nsx-vlan-transportzone-55", "segmentation_id": 55, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf1ef6f0e-1f", "ovs_interfaceid": "f1ef6f0e-1f62-4034-8edb-e8cc1bb98abd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 821.787151] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3e:82:e5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '05b1253d-2b87-4158-9ff1-dafcf829f11f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f1ef6f0e-1f62-4034-8edb-e8cc1bb98abd', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 821.795101] env[63531]: DEBUG oslo.service.loopingcall [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 821.795682] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 821.795976] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d7c2310f-95a9-4036-9a0b-e37d157d8275 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.819252] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 821.819252] env[63531]: value = "task-1117842" [ 821.819252] env[63531]: _type = "Task" [ 821.819252] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.829984] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117842, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.923052] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9e77b61d-e87e-4223-8227-0c685c107554 tempest-ServerActionsV293TestJSON-652999638 tempest-ServerActionsV293TestJSON-652999638-project-member] Lock "410ddf7a-e45b-4769-a3db-5363ce2096a7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.739s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.044614] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1a31986a-b93d-43b8-b2b8-77280719bdff tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Lock "eb6c84d7-e0e6-4999-9018-889b99dc67ad" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 68.037s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.137640] env[63531]: DEBUG nova.compute.manager [req-a897f729-0b19-437d-abf1-52a41dd43855 req-72ed573b-f75e-4fc1-9bed-855be9d7097e service nova] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Received event network-vif-deleted-d99cda7f-476b-4869-8ba7-4a00c03da394 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 822.137935] env[63531]: DEBUG nova.compute.manager [req-a897f729-0b19-437d-abf1-52a41dd43855 req-72ed573b-f75e-4fc1-9bed-855be9d7097e service nova] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Received event network-changed-f1ef6f0e-1f62-4034-8edb-e8cc1bb98abd {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 822.138183] env[63531]: DEBUG nova.compute.manager [req-a897f729-0b19-437d-abf1-52a41dd43855 req-72ed573b-f75e-4fc1-9bed-855be9d7097e service nova] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Refreshing instance network info cache due to event network-changed-f1ef6f0e-1f62-4034-8edb-e8cc1bb98abd. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 822.138419] env[63531]: DEBUG oslo_concurrency.lockutils [req-a897f729-0b19-437d-abf1-52a41dd43855 req-72ed573b-f75e-4fc1-9bed-855be9d7097e service nova] Acquiring lock "refresh_cache-e0074e2e-e8bd-4bcb-9ad9-271a10e079e4" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.138564] env[63531]: DEBUG oslo_concurrency.lockutils [req-a897f729-0b19-437d-abf1-52a41dd43855 req-72ed573b-f75e-4fc1-9bed-855be9d7097e service nova] Acquired lock "refresh_cache-e0074e2e-e8bd-4bcb-9ad9-271a10e079e4" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.138727] env[63531]: DEBUG nova.network.neutron [req-a897f729-0b19-437d-abf1-52a41dd43855 req-72ed573b-f75e-4fc1-9bed-855be9d7097e service nova] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Refreshing network info cache for port f1ef6f0e-1f62-4034-8edb-e8cc1bb98abd {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 822.266015] env[63531]: DEBUG oslo_vmware.api [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1117841, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.268381] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Releasing lock "refresh_cache-5e9042b3-4dc2-4fa3-a664-c4b49a22e400" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.268381] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 822.268381] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f63fd975-3547-4245-9e37-af1ead6f40ba {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.274803] env[63531]: DEBUG oslo_vmware.api [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Waiting for the task: (returnval){ [ 822.274803] env[63531]: value = "task-1117843" [ 822.274803] env[63531]: _type = "Task" [ 822.274803] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.282481] env[63531]: DEBUG oslo_vmware.api [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Task: {'id': task-1117843, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.328735] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117842, 'name': CreateVM_Task, 'duration_secs': 0.41404} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.328904] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 822.329648] env[63531]: DEBUG oslo_concurrency.lockutils [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.329836] env[63531]: DEBUG oslo_concurrency.lockutils [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.330194] env[63531]: DEBUG oslo_concurrency.lockutils [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 822.330455] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-034b07c8-32f0-441d-80b4-85187fac2ece {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.335211] env[63531]: DEBUG oslo_vmware.api [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Waiting for the task: (returnval){ [ 822.335211] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]520c1eec-c8d1-69ab-c384-a3802ce1501a" [ 822.335211] env[63531]: _type = "Task" [ 822.335211] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.342989] env[63531]: DEBUG oslo_vmware.api [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]520c1eec-c8d1-69ab-c384-a3802ce1501a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.408250] env[63531]: DEBUG oslo_vmware.rw_handles [None req-ad241f4e-db19-4dc3-893a-2cb7737a96b1 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5210c391-01c5-4cde-8230-d9ed42991e74/disk-0.vmdk. {{(pid=63531) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 822.409528] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbedf51a-8d2d-4249-b3f0-4f8c19768ffd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.415692] env[63531]: DEBUG oslo_vmware.rw_handles [None req-ad241f4e-db19-4dc3-893a-2cb7737a96b1 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5210c391-01c5-4cde-8230-d9ed42991e74/disk-0.vmdk is in state: ready. {{(pid=63531) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 822.415870] env[63531]: ERROR oslo_vmware.rw_handles [None req-ad241f4e-db19-4dc3-893a-2cb7737a96b1 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5210c391-01c5-4cde-8230-d9ed42991e74/disk-0.vmdk due to incomplete transfer. [ 822.416109] env[63531]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-3fce8d64-c238-4de6-b491-7a94f4e5cc51 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.423065] env[63531]: DEBUG oslo_vmware.rw_handles [None req-ad241f4e-db19-4dc3-893a-2cb7737a96b1 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5210c391-01c5-4cde-8230-d9ed42991e74/disk-0.vmdk. {{(pid=63531) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 822.423277] env[63531]: DEBUG nova.virt.vmwareapi.images [None req-ad241f4e-db19-4dc3-893a-2cb7737a96b1 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Uploaded image 2a1376af-2b52-4754-afef-6403cc9e22a3 to the Glance image server {{(pid=63531) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 822.425494] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad241f4e-db19-4dc3-893a-2cb7737a96b1 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Destroying the VM {{(pid=63531) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 822.425971] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-a8fdc339-003d-4c4e-ae36-3b39d1379169 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.431405] env[63531]: DEBUG oslo_vmware.api [None req-ad241f4e-db19-4dc3-893a-2cb7737a96b1 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Waiting for the task: (returnval){ [ 822.431405] env[63531]: value = "task-1117844" [ 822.431405] env[63531]: _type = "Task" [ 822.431405] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.438660] env[63531]: DEBUG oslo_vmware.api [None req-ad241f4e-db19-4dc3-893a-2cb7737a96b1 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117844, 'name': Destroy_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.470531] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ea661ef7-af43-46e1-812d-b8cc1057ecf2 tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Acquiring lock "eb6c84d7-e0e6-4999-9018-889b99dc67ad" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.470812] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ea661ef7-af43-46e1-812d-b8cc1057ecf2 tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Lock "eb6c84d7-e0e6-4999-9018-889b99dc67ad" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.471030] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ea661ef7-af43-46e1-812d-b8cc1057ecf2 tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Acquiring lock "eb6c84d7-e0e6-4999-9018-889b99dc67ad-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.471216] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ea661ef7-af43-46e1-812d-b8cc1057ecf2 tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Lock "eb6c84d7-e0e6-4999-9018-889b99dc67ad-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.471389] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ea661ef7-af43-46e1-812d-b8cc1057ecf2 tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Lock "eb6c84d7-e0e6-4999-9018-889b99dc67ad-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.474761] env[63531]: INFO nova.compute.manager [None req-ea661ef7-af43-46e1-812d-b8cc1057ecf2 tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] [instance: eb6c84d7-e0e6-4999-9018-889b99dc67ad] Terminating instance [ 822.476651] env[63531]: DEBUG nova.compute.manager [None req-ea661ef7-af43-46e1-812d-b8cc1057ecf2 tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] [instance: eb6c84d7-e0e6-4999-9018-889b99dc67ad] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 822.476860] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-ea661ef7-af43-46e1-812d-b8cc1057ecf2 tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] [instance: eb6c84d7-e0e6-4999-9018-889b99dc67ad] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 822.477750] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af3c9646-4c20-4e3b-8e9b-85b626decea9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.486872] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea661ef7-af43-46e1-812d-b8cc1057ecf2 tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] [instance: eb6c84d7-e0e6-4999-9018-889b99dc67ad] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 822.487203] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dd8dcd3f-cc18-454b-93da-b2dd440c3d32 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.493551] env[63531]: DEBUG oslo_vmware.api [None req-ea661ef7-af43-46e1-812d-b8cc1057ecf2 tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Waiting for the task: (returnval){ [ 822.493551] env[63531]: value = "task-1117845" [ 822.493551] env[63531]: _type = "Task" [ 822.493551] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.503773] env[63531]: DEBUG oslo_vmware.api [None req-ea661ef7-af43-46e1-812d-b8cc1057ecf2 tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Task: {'id': task-1117845, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.547110] env[63531]: DEBUG nova.compute.manager [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: db83adf4-6183-4330-b260-77d1f5daf899] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 822.766301] env[63531]: DEBUG oslo_vmware.api [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1117841, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.780769] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a23d4350-d6b5-4f8b-b684-59af1629d9b6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.786175] env[63531]: DEBUG oslo_vmware.api [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Task: {'id': task-1117843, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.792993] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b094a665-5f96-4452-a9bf-7d949fd75b1a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.835320] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28294540-51b4-4092-b2a9-2d70e6549486 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.851115] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-086e4c7b-0e04-4bc6-ba88-b47cbb17b933 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.856073] env[63531]: DEBUG oslo_vmware.api [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]520c1eec-c8d1-69ab-c384-a3802ce1501a, 'name': SearchDatastore_Task, 'duration_secs': 0.014715} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.856427] env[63531]: DEBUG oslo_concurrency.lockutils [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.856693] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 822.856962] env[63531]: DEBUG oslo_concurrency.lockutils [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.857208] env[63531]: DEBUG oslo_concurrency.lockutils [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.857409] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 822.858129] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c6177a92-f00c-479d-ae85-97e2cde6d96d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.872489] env[63531]: DEBUG nova.compute.provider_tree [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 822.877595] env[63531]: DEBUG nova.network.neutron [req-a897f729-0b19-437d-abf1-52a41dd43855 req-72ed573b-f75e-4fc1-9bed-855be9d7097e service nova] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Updated VIF entry in instance network info cache for port f1ef6f0e-1f62-4034-8edb-e8cc1bb98abd. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 822.878395] env[63531]: DEBUG nova.network.neutron [req-a897f729-0b19-437d-abf1-52a41dd43855 req-72ed573b-f75e-4fc1-9bed-855be9d7097e service nova] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Updating instance_info_cache with network_info: [{"id": "f1ef6f0e-1f62-4034-8edb-e8cc1bb98abd", "address": "fa:16:3e:3e:82:e5", "network": {"id": "dc416058-1c04-47ac-8b73-1557f7c33d1c", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1350689703-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "824ed405cc6e4868a70e82489e05e8ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "05b1253d-2b87-4158-9ff1-dafcf829f11f", "external-id": "nsx-vlan-transportzone-55", "segmentation_id": 55, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf1ef6f0e-1f", "ovs_interfaceid": "f1ef6f0e-1f62-4034-8edb-e8cc1bb98abd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 822.883793] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 822.883793] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 822.885581] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1e8c3c20-9500-4217-beea-e2afbf7e1748 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.892511] env[63531]: DEBUG oslo_vmware.api [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Waiting for the task: (returnval){ [ 822.892511] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52254c04-b3df-455a-7959-fc28e4c74546" [ 822.892511] env[63531]: _type = "Task" [ 822.892511] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.902016] env[63531]: DEBUG oslo_vmware.api [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52254c04-b3df-455a-7959-fc28e4c74546, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.942017] env[63531]: DEBUG oslo_vmware.api [None req-ad241f4e-db19-4dc3-893a-2cb7737a96b1 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117844, 'name': Destroy_Task, 'duration_secs': 0.368839} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.942328] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-ad241f4e-db19-4dc3-893a-2cb7737a96b1 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Destroyed the VM [ 822.942577] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-ad241f4e-db19-4dc3-893a-2cb7737a96b1 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Deleting Snapshot of the VM instance {{(pid=63531) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 822.942839] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-1671d9b6-dd69-4ca3-a2ea-c2badfb5e974 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.949014] env[63531]: DEBUG oslo_vmware.api [None req-ad241f4e-db19-4dc3-893a-2cb7737a96b1 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Waiting for the task: (returnval){ [ 822.949014] env[63531]: value = "task-1117846" [ 822.949014] env[63531]: _type = "Task" [ 822.949014] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.957143] env[63531]: DEBUG oslo_vmware.api [None req-ad241f4e-db19-4dc3-893a-2cb7737a96b1 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117846, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.006066] env[63531]: DEBUG oslo_vmware.api [None req-ea661ef7-af43-46e1-812d-b8cc1057ecf2 tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Task: {'id': task-1117845, 'name': PowerOffVM_Task, 'duration_secs': 0.397696} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.006184] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea661ef7-af43-46e1-812d-b8cc1057ecf2 tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] [instance: eb6c84d7-e0e6-4999-9018-889b99dc67ad] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 823.006409] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-ea661ef7-af43-46e1-812d-b8cc1057ecf2 tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] [instance: eb6c84d7-e0e6-4999-9018-889b99dc67ad] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 823.006682] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-262f5691-68d9-41c4-845f-92b8e4b4b0c3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.068803] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 823.080656] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-ea661ef7-af43-46e1-812d-b8cc1057ecf2 tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] [instance: eb6c84d7-e0e6-4999-9018-889b99dc67ad] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 823.081030] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-ea661ef7-af43-46e1-812d-b8cc1057ecf2 tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] [instance: eb6c84d7-e0e6-4999-9018-889b99dc67ad] Deleting contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 823.081194] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea661ef7-af43-46e1-812d-b8cc1057ecf2 tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Deleting the datastore file [datastore2] eb6c84d7-e0e6-4999-9018-889b99dc67ad {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 823.081517] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4b24f7ca-eef2-40a1-a1ab-bcdb69a4b2a6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.087669] env[63531]: DEBUG oslo_vmware.api [None req-ea661ef7-af43-46e1-812d-b8cc1057ecf2 tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Waiting for the task: (returnval){ [ 823.087669] env[63531]: value = "task-1117848" [ 823.087669] env[63531]: _type = "Task" [ 823.087669] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.097312] env[63531]: DEBUG oslo_vmware.api [None req-ea661ef7-af43-46e1-812d-b8cc1057ecf2 tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Task: {'id': task-1117848, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.268578] env[63531]: DEBUG oslo_vmware.api [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1117841, 'name': ReconfigVM_Task, 'duration_secs': 1.337942} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.268578] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Reconfigured VM instance instance-0000003c to attach disk [datastore2] f46d8a85-6ab7-436a-bc5e-40d0f80974e5/f46d8a85-6ab7-436a-bc5e-40d0f80974e5.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 823.269219] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5dc7e75a-69c4-450d-a00a-0897314aa71e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.275968] env[63531]: DEBUG oslo_vmware.api [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Waiting for the task: (returnval){ [ 823.275968] env[63531]: value = "task-1117849" [ 823.275968] env[63531]: _type = "Task" [ 823.275968] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.288788] env[63531]: DEBUG oslo_vmware.api [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1117849, 'name': Rename_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.291913] env[63531]: DEBUG oslo_vmware.api [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Task: {'id': task-1117843, 'name': PowerOffVM_Task, 'duration_secs': 0.616187} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.292187] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 823.292953] env[63531]: DEBUG nova.virt.hardware [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:17:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='86bdb6e9-50cf-4ad6-bad7-f8cb5dc21d1b',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-672645615',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 823.293208] env[63531]: DEBUG nova.virt.hardware [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 823.293390] env[63531]: DEBUG nova.virt.hardware [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 823.293599] env[63531]: DEBUG nova.virt.hardware [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 823.293757] env[63531]: DEBUG nova.virt.hardware [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 823.293925] env[63531]: DEBUG nova.virt.hardware [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 823.294162] env[63531]: DEBUG nova.virt.hardware [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 823.294347] env[63531]: DEBUG nova.virt.hardware [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 823.294535] env[63531]: DEBUG nova.virt.hardware [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 823.294718] env[63531]: DEBUG nova.virt.hardware [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 823.294928] env[63531]: DEBUG nova.virt.hardware [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 823.300181] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bb4aee41-2019-4d13-9815-e02d33a24efa {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.315228] env[63531]: DEBUG oslo_vmware.api [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Waiting for the task: (returnval){ [ 823.315228] env[63531]: value = "task-1117850" [ 823.315228] env[63531]: _type = "Task" [ 823.315228] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.324051] env[63531]: DEBUG oslo_vmware.api [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Task: {'id': task-1117850, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.376622] env[63531]: DEBUG nova.scheduler.client.report [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 823.381905] env[63531]: DEBUG oslo_concurrency.lockutils [req-a897f729-0b19-437d-abf1-52a41dd43855 req-72ed573b-f75e-4fc1-9bed-855be9d7097e service nova] Releasing lock "refresh_cache-e0074e2e-e8bd-4bcb-9ad9-271a10e079e4" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.404365] env[63531]: DEBUG oslo_vmware.api [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52254c04-b3df-455a-7959-fc28e4c74546, 'name': SearchDatastore_Task, 'duration_secs': 0.043517} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.405337] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-802c8e78-d415-4665-a00a-5263de1e047c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.411706] env[63531]: DEBUG oslo_vmware.api [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Waiting for the task: (returnval){ [ 823.411706] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]527e6673-2cd5-7f18-3d6d-e7a218565e2b" [ 823.411706] env[63531]: _type = "Task" [ 823.411706] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.424429] env[63531]: DEBUG oslo_vmware.api [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]527e6673-2cd5-7f18-3d6d-e7a218565e2b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.460479] env[63531]: DEBUG oslo_vmware.api [None req-ad241f4e-db19-4dc3-893a-2cb7737a96b1 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117846, 'name': RemoveSnapshot_Task} progress is 29%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.599570] env[63531]: DEBUG oslo_vmware.api [None req-ea661ef7-af43-46e1-812d-b8cc1057ecf2 tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Task: {'id': task-1117848, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.341006} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.599969] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea661ef7-af43-46e1-812d-b8cc1057ecf2 tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 823.600385] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-ea661ef7-af43-46e1-812d-b8cc1057ecf2 tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] [instance: eb6c84d7-e0e6-4999-9018-889b99dc67ad] Deleted contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 823.600666] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-ea661ef7-af43-46e1-812d-b8cc1057ecf2 tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] [instance: eb6c84d7-e0e6-4999-9018-889b99dc67ad] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 823.600985] env[63531]: INFO nova.compute.manager [None req-ea661ef7-af43-46e1-812d-b8cc1057ecf2 tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] [instance: eb6c84d7-e0e6-4999-9018-889b99dc67ad] Took 1.12 seconds to destroy the instance on the hypervisor. [ 823.601371] env[63531]: DEBUG oslo.service.loopingcall [None req-ea661ef7-af43-46e1-812d-b8cc1057ecf2 tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 823.601653] env[63531]: DEBUG nova.compute.manager [-] [instance: eb6c84d7-e0e6-4999-9018-889b99dc67ad] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 823.601834] env[63531]: DEBUG nova.network.neutron [-] [instance: eb6c84d7-e0e6-4999-9018-889b99dc67ad] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 823.788762] env[63531]: DEBUG oslo_vmware.api [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1117849, 'name': Rename_Task, 'duration_secs': 0.150697} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.789145] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 823.789476] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fc1bb0eb-28b8-4f52-a804-68994e31e65d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.796100] env[63531]: DEBUG oslo_vmware.api [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Waiting for the task: (returnval){ [ 823.796100] env[63531]: value = "task-1117851" [ 823.796100] env[63531]: _type = "Task" [ 823.796100] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.804582] env[63531]: DEBUG oslo_vmware.api [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1117851, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.826374] env[63531]: DEBUG oslo_vmware.api [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Task: {'id': task-1117850, 'name': ReconfigVM_Task, 'duration_secs': 0.149824} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.828742] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7442ff38-2877-43ab-b89b-5fcd36860284 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.848839] env[63531]: DEBUG nova.virt.hardware [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:17:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='86bdb6e9-50cf-4ad6-bad7-f8cb5dc21d1b',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-672645615',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 823.849158] env[63531]: DEBUG nova.virt.hardware [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 823.849354] env[63531]: DEBUG nova.virt.hardware [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 823.849610] env[63531]: DEBUG nova.virt.hardware [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 823.849786] env[63531]: DEBUG nova.virt.hardware [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 823.849942] env[63531]: DEBUG nova.virt.hardware [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 823.850181] env[63531]: DEBUG nova.virt.hardware [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 823.850377] env[63531]: DEBUG nova.virt.hardware [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 823.850567] env[63531]: DEBUG nova.virt.hardware [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 823.850752] env[63531]: DEBUG nova.virt.hardware [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 823.850962] env[63531]: DEBUG nova.virt.hardware [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 823.851832] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5323d066-7ad6-456d-9a9c-3b6c68593d54 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.857723] env[63531]: DEBUG oslo_vmware.api [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Waiting for the task: (returnval){ [ 823.857723] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52d9f62f-7096-af76-4876-3f8bb98c85da" [ 823.857723] env[63531]: _type = "Task" [ 823.857723] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.866566] env[63531]: DEBUG oslo_vmware.api [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52d9f62f-7096-af76-4876-3f8bb98c85da, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.885042] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.478s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.885322] env[63531]: DEBUG nova.compute.manager [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] [instance: 43c29443-1210-4ee2-95c7-1257de308287] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 823.888084] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 32.977s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.925192] env[63531]: DEBUG oslo_vmware.api [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]527e6673-2cd5-7f18-3d6d-e7a218565e2b, 'name': SearchDatastore_Task, 'duration_secs': 0.02336} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.925459] env[63531]: DEBUG oslo_concurrency.lockutils [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.925779] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] e0074e2e-e8bd-4bcb-9ad9-271a10e079e4/e0074e2e-e8bd-4bcb-9ad9-271a10e079e4.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 823.926111] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d196ffab-ce9e-4528-8fbb-66ea97d658f3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.934642] env[63531]: DEBUG oslo_vmware.api [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Waiting for the task: (returnval){ [ 823.934642] env[63531]: value = "task-1117852" [ 823.934642] env[63531]: _type = "Task" [ 823.934642] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.944677] env[63531]: DEBUG oslo_vmware.api [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117852, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.962133] env[63531]: DEBUG oslo_vmware.api [None req-ad241f4e-db19-4dc3-893a-2cb7737a96b1 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117846, 'name': RemoveSnapshot_Task, 'duration_secs': 0.882242} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.962665] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-ad241f4e-db19-4dc3-893a-2cb7737a96b1 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Deleted Snapshot of the VM instance {{(pid=63531) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 823.962775] env[63531]: INFO nova.compute.manager [None req-ad241f4e-db19-4dc3-893a-2cb7737a96b1 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Took 17.52 seconds to snapshot the instance on the hypervisor. [ 824.161146] env[63531]: DEBUG nova.compute.manager [req-c32f2b49-1a32-4a1c-a522-df294922a02d req-89a1703a-13e7-49b3-baf8-6e61ffbcafa4 service nova] [instance: eb6c84d7-e0e6-4999-9018-889b99dc67ad] Received event network-vif-deleted-01bd4815-38ea-44f3-b7d3-c24fef3fd863 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 824.161407] env[63531]: INFO nova.compute.manager [req-c32f2b49-1a32-4a1c-a522-df294922a02d req-89a1703a-13e7-49b3-baf8-6e61ffbcafa4 service nova] [instance: eb6c84d7-e0e6-4999-9018-889b99dc67ad] Neutron deleted interface 01bd4815-38ea-44f3-b7d3-c24fef3fd863; detaching it from the instance and deleting it from the info cache [ 824.161544] env[63531]: DEBUG nova.network.neutron [req-c32f2b49-1a32-4a1c-a522-df294922a02d req-89a1703a-13e7-49b3-baf8-6e61ffbcafa4 service nova] [instance: eb6c84d7-e0e6-4999-9018-889b99dc67ad] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.307166] env[63531]: DEBUG oslo_vmware.api [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1117851, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.367799] env[63531]: DEBUG oslo_vmware.api [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52d9f62f-7096-af76-4876-3f8bb98c85da, 'name': SearchDatastore_Task, 'duration_secs': 0.007582} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.373297] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Reconfiguring VM instance instance-00000027 to detach disk 2000 {{(pid=63531) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 824.373671] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c686c55b-8c43-4f7a-bcbf-4358bd30efd5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.392820] env[63531]: DEBUG nova.compute.utils [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 824.394313] env[63531]: DEBUG oslo_vmware.api [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Waiting for the task: (returnval){ [ 824.394313] env[63531]: value = "task-1117853" [ 824.394313] env[63531]: _type = "Task" [ 824.394313] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.394789] env[63531]: DEBUG nova.compute.manager [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] [instance: 43c29443-1210-4ee2-95c7-1257de308287] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 824.395182] env[63531]: DEBUG nova.network.neutron [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] [instance: 43c29443-1210-4ee2-95c7-1257de308287] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 824.415934] env[63531]: DEBUG oslo_vmware.api [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Task: {'id': task-1117853, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.444380] env[63531]: DEBUG oslo_vmware.api [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117852, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.458461] env[63531]: DEBUG nova.policy [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1e74179e49b64e80b33a0afff66091ea', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8755c58b4afa4fb08a667efc33734ce6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 824.459831] env[63531]: DEBUG nova.network.neutron [-] [instance: eb6c84d7-e0e6-4999-9018-889b99dc67ad] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.664176] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d737702a-83c4-4140-94bb-391f9edeef72 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.672778] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b3d2e5e-4101-48f6-a89b-b24d92e2609e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.705148] env[63531]: DEBUG nova.compute.manager [req-c32f2b49-1a32-4a1c-a522-df294922a02d req-89a1703a-13e7-49b3-baf8-6e61ffbcafa4 service nova] [instance: eb6c84d7-e0e6-4999-9018-889b99dc67ad] Detach interface failed, port_id=01bd4815-38ea-44f3-b7d3-c24fef3fd863, reason: Instance eb6c84d7-e0e6-4999-9018-889b99dc67ad could not be found. {{(pid=63531) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 824.808480] env[63531]: DEBUG oslo_vmware.api [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1117851, 'name': PowerOnVM_Task, 'duration_secs': 0.825894} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.808928] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 824.809172] env[63531]: INFO nova.compute.manager [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Took 10.85 seconds to spawn the instance on the hypervisor. [ 824.809363] env[63531]: DEBUG nova.compute.manager [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 824.810201] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38e2fdda-1e29-4c59-bd85-834872971b53 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.885674] env[63531]: DEBUG nova.network.neutron [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] [instance: 43c29443-1210-4ee2-95c7-1257de308287] Successfully created port: 73fd445a-3029-48b0-b3d7-7246dc21081b {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 824.895707] env[63531]: DEBUG nova.compute.manager [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] [instance: 43c29443-1210-4ee2-95c7-1257de308287] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 824.923091] env[63531]: DEBUG oslo_vmware.api [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Task: {'id': task-1117853, 'name': ReconfigVM_Task, 'duration_secs': 0.223112} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.923091] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Reconfigured VM instance instance-00000027 to detach disk 2000 {{(pid=63531) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 824.923091] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8ba6270-336f-4aef-b179-89c16a408110 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.946574] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Reconfiguring VM instance instance-00000027 to attach disk [datastore1] 5e9042b3-4dc2-4fa3-a664-c4b49a22e400/5e9042b3-4dc2-4fa3-a664-c4b49a22e400.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 824.948129] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 1ca7b851-2a06-4181-8271-58aafcd322d6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 824.948129] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance e7fb00a3-0d2e-4f54-950d-337307112d7a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 824.948129] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 29dd6de3-2765-408c-acc0-da47e5e0a977 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 824.948321] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 8dec0cac-defc-42f2-ab56-6b3ae60ad858 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 824.948321] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 824.948445] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 64fb5831-0789-482b-bb98-67b29868c4c9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 824.948763] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 0197e03d-6c36-4e73-9472-8e3319eb89ce actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 824.948763] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 824.948763] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 9a3fad46-a15a-451c-bdab-a3c8cc8add07 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 824.948884] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 48043b5d-c0d3-4e8a-8d28-5a408d800e11 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 824.948996] env[63531]: WARNING nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance caf0dee8-cb5a-402b-b533-22a590433f54 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 824.949125] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 824.949238] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 5e9042b3-4dc2-4fa3-a664-c4b49a22e400 actively managed on this compute host and has allocations in placement: {'resources': {'VCPU': 1, 'MEMORY_MB': 192, 'DISK_GB': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 824.949359] env[63531]: WARNING nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 531e1852-0c67-4b4d-a0ca-749ac438e79a is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 824.949471] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance eb6c84d7-e0e6-4999-9018-889b99dc67ad actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 824.949698] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance f46d8a85-6ab7-436a-bc5e-40d0f80974e5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 824.949698] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance e0074e2e-e8bd-4bcb-9ad9-271a10e079e4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 824.949793] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 43c29443-1210-4ee2-95c7-1257de308287 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 824.954652] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1f6b2d84-b973-48dd-8095-ea9a7d565313 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.967545] env[63531]: INFO nova.compute.manager [-] [instance: eb6c84d7-e0e6-4999-9018-889b99dc67ad] Took 1.37 seconds to deallocate network for instance. [ 824.976356] env[63531]: DEBUG oslo_vmware.api [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117852, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.564091} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.977845] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] e0074e2e-e8bd-4bcb-9ad9-271a10e079e4/e0074e2e-e8bd-4bcb-9ad9-271a10e079e4.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 824.978093] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 824.978449] env[63531]: DEBUG oslo_vmware.api [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Waiting for the task: (returnval){ [ 824.978449] env[63531]: value = "task-1117854" [ 824.978449] env[63531]: _type = "Task" [ 824.978449] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.978872] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2cef06c1-dde4-410b-844e-d1f8d6133ff6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.990989] env[63531]: DEBUG oslo_vmware.api [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Task: {'id': task-1117854, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.992357] env[63531]: DEBUG oslo_vmware.api [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Waiting for the task: (returnval){ [ 824.992357] env[63531]: value = "task-1117855" [ 824.992357] env[63531]: _type = "Task" [ 824.992357] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.000832] env[63531]: DEBUG oslo_vmware.api [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117855, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.332792] env[63531]: INFO nova.compute.manager [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Took 41.16 seconds to build instance. [ 825.468224] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 45b05c54-6865-47a1-ba93-90ad3e1ba07e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 825.478687] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ea661ef7-af43-46e1-812d-b8cc1057ecf2 tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.490477] env[63531]: DEBUG oslo_vmware.api [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Task: {'id': task-1117854, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.500406] env[63531]: DEBUG oslo_vmware.api [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117855, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.310464} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.500668] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 825.501417] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c2e7658-c98a-4022-a28b-8750e2a39cfb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.522662] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Reconfiguring VM instance instance-0000003d to attach disk [datastore2] e0074e2e-e8bd-4bcb-9ad9-271a10e079e4/e0074e2e-e8bd-4bcb-9ad9-271a10e079e4.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 825.523421] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ff95fd18-8877-4b10-a0f6-794976136ea9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.541912] env[63531]: DEBUG oslo_vmware.api [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Waiting for the task: (returnval){ [ 825.541912] env[63531]: value = "task-1117856" [ 825.541912] env[63531]: _type = "Task" [ 825.541912] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.549696] env[63531]: DEBUG oslo_vmware.api [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117856, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.837910] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6b69b22b-667b-42e9-9502-10108e0decdc tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lock "f46d8a85-6ab7-436a-bc5e-40d0f80974e5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 70.618s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.904150] env[63531]: DEBUG nova.compute.manager [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] [instance: 43c29443-1210-4ee2-95c7-1257de308287] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 825.926359] env[63531]: DEBUG nova.virt.hardware [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 825.926640] env[63531]: DEBUG nova.virt.hardware [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 825.926818] env[63531]: DEBUG nova.virt.hardware [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 825.927036] env[63531]: DEBUG nova.virt.hardware [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 825.927270] env[63531]: DEBUG nova.virt.hardware [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 825.927480] env[63531]: DEBUG nova.virt.hardware [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 825.927725] env[63531]: DEBUG nova.virt.hardware [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 825.927910] env[63531]: DEBUG nova.virt.hardware [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 825.928113] env[63531]: DEBUG nova.virt.hardware [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 825.928296] env[63531]: DEBUG nova.virt.hardware [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 825.928481] env[63531]: DEBUG nova.virt.hardware [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 825.929346] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7569430e-3194-46ac-8661-cd4a41c32f37 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.937895] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ae2d0fa-1fd9-41e9-838c-396093983078 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.970729] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 825.990696] env[63531]: DEBUG oslo_vmware.api [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Task: {'id': task-1117854, 'name': ReconfigVM_Task, 'duration_secs': 0.971625} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.991013] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Reconfigured VM instance instance-00000027 to attach disk [datastore1] 5e9042b3-4dc2-4fa3-a664-c4b49a22e400/5e9042b3-4dc2-4fa3-a664-c4b49a22e400.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 825.991843] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ed73ff8-f7ef-4201-83ed-f2c83bc99af7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.012130] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f57d702-1d5f-45ab-92ae-c33eada673c4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.030810] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1e1c7c4-aec3-4049-8f67-3ed79248af36 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.052032] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c39b970-4bda-4253-b5d1-6a3b03d7c502 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.059626] env[63531]: DEBUG oslo_vmware.api [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117856, 'name': ReconfigVM_Task, 'duration_secs': 0.307117} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.061902] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Reconfigured VM instance instance-0000003d to attach disk [datastore2] e0074e2e-e8bd-4bcb-9ad9-271a10e079e4/e0074e2e-e8bd-4bcb-9ad9-271a10e079e4.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 826.062622] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 826.063150] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0dcba826-cd6a-44c9-a68d-a4d1a23b5286 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.064512] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-35035278-a4bb-4291-88a5-27e0eb60aa76 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.070119] env[63531]: DEBUG oslo_vmware.api [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Waiting for the task: (returnval){ [ 826.070119] env[63531]: value = "task-1117857" [ 826.070119] env[63531]: _type = "Task" [ 826.070119] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.071263] env[63531]: DEBUG oslo_vmware.api [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Waiting for the task: (returnval){ [ 826.071263] env[63531]: value = "task-1117858" [ 826.071263] env[63531]: _type = "Task" [ 826.071263] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.082224] env[63531]: DEBUG oslo_vmware.api [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Task: {'id': task-1117858, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.085270] env[63531]: DEBUG oslo_vmware.api [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117857, 'name': Rename_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.260743] env[63531]: DEBUG nova.compute.manager [None req-13b8019c-692f-4097-a7ee-b3251d4a991c tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 826.262080] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e332f070-048f-469b-8ed9-c8fc9d5ba448 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.341204] env[63531]: DEBUG nova.compute.manager [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 826.473489] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 3eeaa6f9-60d9-43fe-a438-fbf2ceececae has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 826.583195] env[63531]: DEBUG oslo_vmware.api [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117857, 'name': Rename_Task, 'duration_secs': 0.384508} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.586099] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 826.586370] env[63531]: DEBUG oslo_vmware.api [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Task: {'id': task-1117858, 'name': PowerOnVM_Task, 'duration_secs': 0.452865} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.586743] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7fe9c916-24b5-4ac6-9725-89291d96d8c4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.588046] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 826.595800] env[63531]: DEBUG oslo_vmware.api [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Waiting for the task: (returnval){ [ 826.595800] env[63531]: value = "task-1117859" [ 826.595800] env[63531]: _type = "Task" [ 826.595800] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.605526] env[63531]: DEBUG oslo_vmware.api [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117859, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.776085] env[63531]: INFO nova.compute.manager [None req-13b8019c-692f-4097-a7ee-b3251d4a991c tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] instance snapshotting [ 826.779092] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-141d43e7-2ee2-4698-92c8-14f1392da529 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.798585] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-628ee64e-efdc-4ceb-b567-524bc0a229f3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.869446] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.977374] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 09feb12a-da9f-4bb7-959d-84c5c7c58e99 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 827.105973] env[63531]: DEBUG oslo_vmware.api [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117859, 'name': PowerOnVM_Task, 'duration_secs': 0.446002} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.106541] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 827.106541] env[63531]: INFO nova.compute.manager [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Took 8.92 seconds to spawn the instance on the hypervisor. [ 827.106718] env[63531]: DEBUG nova.compute.manager [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 827.107432] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-239f5542-6aff-4f67-91f4-96d3c3284976 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.203082] env[63531]: DEBUG nova.compute.manager [req-5a87e6a7-1086-436e-b8ba-c969fb7e0374 req-a1af74a9-e838-4cce-8e85-692ca102113d service nova] [instance: 43c29443-1210-4ee2-95c7-1257de308287] Received event network-vif-plugged-73fd445a-3029-48b0-b3d7-7246dc21081b {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 827.203402] env[63531]: DEBUG oslo_concurrency.lockutils [req-5a87e6a7-1086-436e-b8ba-c969fb7e0374 req-a1af74a9-e838-4cce-8e85-692ca102113d service nova] Acquiring lock "43c29443-1210-4ee2-95c7-1257de308287-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.204013] env[63531]: DEBUG oslo_concurrency.lockutils [req-5a87e6a7-1086-436e-b8ba-c969fb7e0374 req-a1af74a9-e838-4cce-8e85-692ca102113d service nova] Lock "43c29443-1210-4ee2-95c7-1257de308287-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.204212] env[63531]: DEBUG oslo_concurrency.lockutils [req-5a87e6a7-1086-436e-b8ba-c969fb7e0374 req-a1af74a9-e838-4cce-8e85-692ca102113d service nova] Lock "43c29443-1210-4ee2-95c7-1257de308287-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.204381] env[63531]: DEBUG nova.compute.manager [req-5a87e6a7-1086-436e-b8ba-c969fb7e0374 req-a1af74a9-e838-4cce-8e85-692ca102113d service nova] [instance: 43c29443-1210-4ee2-95c7-1257de308287] No waiting events found dispatching network-vif-plugged-73fd445a-3029-48b0-b3d7-7246dc21081b {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 827.204557] env[63531]: WARNING nova.compute.manager [req-5a87e6a7-1086-436e-b8ba-c969fb7e0374 req-a1af74a9-e838-4cce-8e85-692ca102113d service nova] [instance: 43c29443-1210-4ee2-95c7-1257de308287] Received unexpected event network-vif-plugged-73fd445a-3029-48b0-b3d7-7246dc21081b for instance with vm_state building and task_state spawning. [ 827.276570] env[63531]: DEBUG nova.compute.manager [req-b18a3468-f85c-44fd-bf7d-77e4ba22d160 req-be054d0b-7fda-4f03-a74a-76c0fad36a33 service nova] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Received event network-changed-a3fd640b-c47b-4501-ac04-f90f6af8b076 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 827.276808] env[63531]: DEBUG nova.compute.manager [req-b18a3468-f85c-44fd-bf7d-77e4ba22d160 req-be054d0b-7fda-4f03-a74a-76c0fad36a33 service nova] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Refreshing instance network info cache due to event network-changed-a3fd640b-c47b-4501-ac04-f90f6af8b076. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 827.277040] env[63531]: DEBUG oslo_concurrency.lockutils [req-b18a3468-f85c-44fd-bf7d-77e4ba22d160 req-be054d0b-7fda-4f03-a74a-76c0fad36a33 service nova] Acquiring lock "refresh_cache-f46d8a85-6ab7-436a-bc5e-40d0f80974e5" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 827.277223] env[63531]: DEBUG oslo_concurrency.lockutils [req-b18a3468-f85c-44fd-bf7d-77e4ba22d160 req-be054d0b-7fda-4f03-a74a-76c0fad36a33 service nova] Acquired lock "refresh_cache-f46d8a85-6ab7-436a-bc5e-40d0f80974e5" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.277392] env[63531]: DEBUG nova.network.neutron [req-b18a3468-f85c-44fd-bf7d-77e4ba22d160 req-be054d0b-7fda-4f03-a74a-76c0fad36a33 service nova] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Refreshing network info cache for port a3fd640b-c47b-4501-ac04-f90f6af8b076 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 827.309032] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-13b8019c-692f-4097-a7ee-b3251d4a991c tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Creating Snapshot of the VM instance {{(pid=63531) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 827.309209] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-2b2bf8ef-56bd-450f-8336-1b475453ec2e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.316931] env[63531]: DEBUG oslo_vmware.api [None req-13b8019c-692f-4097-a7ee-b3251d4a991c tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Waiting for the task: (returnval){ [ 827.316931] env[63531]: value = "task-1117860" [ 827.316931] env[63531]: _type = "Task" [ 827.316931] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.323008] env[63531]: DEBUG nova.objects.instance [None req-3ae55c07-6554-4c57-846d-88603c4901e5 tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Lazy-loading 'flavor' on Instance uuid 0197e03d-6c36-4e73-9472-8e3319eb89ce {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 827.329265] env[63531]: DEBUG oslo_vmware.api [None req-13b8019c-692f-4097-a7ee-b3251d4a991c tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117860, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.481013] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 2b0e4285-310a-439c-bc50-000a7c5ef7f9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 827.598887] env[63531]: INFO nova.compute.manager [None req-b400dd08-6305-453b-83ef-1ae4e2d27fdf tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Updating instance to original state: 'active' [ 827.610551] env[63531]: DEBUG nova.network.neutron [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] [instance: 43c29443-1210-4ee2-95c7-1257de308287] Successfully updated port: 73fd445a-3029-48b0-b3d7-7246dc21081b {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 827.629337] env[63531]: INFO nova.compute.manager [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Took 42.08 seconds to build instance. [ 827.829643] env[63531]: DEBUG oslo_vmware.api [None req-13b8019c-692f-4097-a7ee-b3251d4a991c tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117860, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.830458] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3ae55c07-6554-4c57-846d-88603c4901e5 tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Acquiring lock "refresh_cache-0197e03d-6c36-4e73-9472-8e3319eb89ce" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 827.830458] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3ae55c07-6554-4c57-846d-88603c4901e5 tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Acquired lock "refresh_cache-0197e03d-6c36-4e73-9472-8e3319eb89ce" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.988864] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 828.113000] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Acquiring lock "refresh_cache-43c29443-1210-4ee2-95c7-1257de308287" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.113163] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Acquired lock "refresh_cache-43c29443-1210-4ee2-95c7-1257de308287" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.113326] env[63531]: DEBUG nova.network.neutron [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] [instance: 43c29443-1210-4ee2-95c7-1257de308287] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 828.122775] env[63531]: DEBUG nova.network.neutron [req-b18a3468-f85c-44fd-bf7d-77e4ba22d160 req-be054d0b-7fda-4f03-a74a-76c0fad36a33 service nova] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Updated VIF entry in instance network info cache for port a3fd640b-c47b-4501-ac04-f90f6af8b076. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 828.123185] env[63531]: DEBUG nova.network.neutron [req-b18a3468-f85c-44fd-bf7d-77e4ba22d160 req-be054d0b-7fda-4f03-a74a-76c0fad36a33 service nova] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Updating instance_info_cache with network_info: [{"id": "a3fd640b-c47b-4501-ac04-f90f6af8b076", "address": "fa:16:3e:27:ca:51", "network": {"id": "4dd26082-3411-4d99-a78b-c76c15703506", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1190890625-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.196", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ee5fd84e1da04fd984f81578e4085850", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6046aec4-feda-4ef9-bf4a-800de1e0cd3b", "external-id": "nsx-vlan-transportzone-903", "segmentation_id": 903, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3fd640b-c4", "ovs_interfaceid": "a3fd640b-c47b-4501-ac04-f90f6af8b076", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 828.130762] env[63531]: DEBUG oslo_concurrency.lockutils [None req-694ad79a-6f4c-40d7-9053-525fb61bc1f4 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Lock "e0074e2e-e8bd-4bcb-9ad9-271a10e079e4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 70.131s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.229277] env[63531]: DEBUG nova.network.neutron [None req-3ae55c07-6554-4c57-846d-88603c4901e5 tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 828.327820] env[63531]: DEBUG oslo_vmware.api [None req-13b8019c-692f-4097-a7ee-b3251d4a991c tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117860, 'name': CreateSnapshot_Task, 'duration_secs': 0.578847} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.329215] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-13b8019c-692f-4097-a7ee-b3251d4a991c tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Created Snapshot of the VM instance {{(pid=63531) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 828.330362] env[63531]: DEBUG nova.compute.manager [req-ea2a35b8-a2da-44b0-8942-9a1e5b913e35 req-6202a6df-9f65-47b7-87db-2e49af35b64b service nova] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Received event network-changed-b4141cf4-dab5-4033-8cc5-06966340323b {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 828.330542] env[63531]: DEBUG nova.compute.manager [req-ea2a35b8-a2da-44b0-8942-9a1e5b913e35 req-6202a6df-9f65-47b7-87db-2e49af35b64b service nova] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Refreshing instance network info cache due to event network-changed-b4141cf4-dab5-4033-8cc5-06966340323b. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 828.330733] env[63531]: DEBUG oslo_concurrency.lockutils [req-ea2a35b8-a2da-44b0-8942-9a1e5b913e35 req-6202a6df-9f65-47b7-87db-2e49af35b64b service nova] Acquiring lock "refresh_cache-0197e03d-6c36-4e73-9472-8e3319eb89ce" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.331451] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ea1a557-0325-4c0a-8362-4f5066619540 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.492654] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance db83adf4-6183-4330-b260-77d1f5daf899 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 828.627412] env[63531]: DEBUG oslo_concurrency.lockutils [req-b18a3468-f85c-44fd-bf7d-77e4ba22d160 req-be054d0b-7fda-4f03-a74a-76c0fad36a33 service nova] Releasing lock "refresh_cache-f46d8a85-6ab7-436a-bc5e-40d0f80974e5" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 828.633164] env[63531]: DEBUG nova.compute.manager [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 828.647858] env[63531]: DEBUG nova.network.neutron [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] [instance: 43c29443-1210-4ee2-95c7-1257de308287] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 828.702558] env[63531]: INFO nova.compute.manager [None req-578f74de-4af2-4ff1-9a49-a39d5b6c11ae tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Rescuing [ 828.702828] env[63531]: DEBUG oslo_concurrency.lockutils [None req-578f74de-4af2-4ff1-9a49-a39d5b6c11ae tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Acquiring lock "refresh_cache-e0074e2e-e8bd-4bcb-9ad9-271a10e079e4" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.702987] env[63531]: DEBUG oslo_concurrency.lockutils [None req-578f74de-4af2-4ff1-9a49-a39d5b6c11ae tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Acquired lock "refresh_cache-e0074e2e-e8bd-4bcb-9ad9-271a10e079e4" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.703179] env[63531]: DEBUG nova.network.neutron [None req-578f74de-4af2-4ff1-9a49-a39d5b6c11ae tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 828.810587] env[63531]: DEBUG nova.network.neutron [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] [instance: 43c29443-1210-4ee2-95c7-1257de308287] Updating instance_info_cache with network_info: [{"id": "73fd445a-3029-48b0-b3d7-7246dc21081b", "address": "fa:16:3e:bf:a4:b6", "network": {"id": "e01dcba8-0a6e-4458-aab6-e97c2c843bea", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-516244587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8755c58b4afa4fb08a667efc33734ce6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b7bf7d4-8e0c-4cee-84ba-244e73ef6379", "external-id": "nsx-vlan-transportzone-423", "segmentation_id": 423, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap73fd445a-30", "ovs_interfaceid": "73fd445a-3029-48b0-b3d7-7246dc21081b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 828.853901] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-13b8019c-692f-4097-a7ee-b3251d4a991c tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Creating linked-clone VM from snapshot {{(pid=63531) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 828.854836] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-e691c62f-09dc-4761-afd4-bdd0e37ce4aa {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.865842] env[63531]: DEBUG oslo_vmware.api [None req-13b8019c-692f-4097-a7ee-b3251d4a991c tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Waiting for the task: (returnval){ [ 828.865842] env[63531]: value = "task-1117861" [ 828.865842] env[63531]: _type = "Task" [ 828.865842] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.875707] env[63531]: DEBUG oslo_vmware.api [None req-13b8019c-692f-4097-a7ee-b3251d4a991c tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117861, 'name': CloneVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.996289] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 829.005813] env[63531]: DEBUG oslo_concurrency.lockutils [None req-932f1c32-d337-45aa-af0e-9ce7ab65093a tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Acquiring lock "5e9042b3-4dc2-4fa3-a664-c4b49a22e400" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.006097] env[63531]: DEBUG oslo_concurrency.lockutils [None req-932f1c32-d337-45aa-af0e-9ce7ab65093a tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Lock "5e9042b3-4dc2-4fa3-a664-c4b49a22e400" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.006349] env[63531]: DEBUG oslo_concurrency.lockutils [None req-932f1c32-d337-45aa-af0e-9ce7ab65093a tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Acquiring lock "5e9042b3-4dc2-4fa3-a664-c4b49a22e400-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.006499] env[63531]: DEBUG oslo_concurrency.lockutils [None req-932f1c32-d337-45aa-af0e-9ce7ab65093a tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Lock "5e9042b3-4dc2-4fa3-a664-c4b49a22e400-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.006676] env[63531]: DEBUG oslo_concurrency.lockutils [None req-932f1c32-d337-45aa-af0e-9ce7ab65093a tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Lock "5e9042b3-4dc2-4fa3-a664-c4b49a22e400-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.009167] env[63531]: INFO nova.compute.manager [None req-932f1c32-d337-45aa-af0e-9ce7ab65093a tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Terminating instance [ 829.011256] env[63531]: DEBUG nova.compute.manager [None req-932f1c32-d337-45aa-af0e-9ce7ab65093a tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 829.011481] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-932f1c32-d337-45aa-af0e-9ce7ab65093a tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 829.012320] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ac262dc-7877-4ad5-9ada-0d6cfee19969 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.022215] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-932f1c32-d337-45aa-af0e-9ce7ab65093a tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 829.022599] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0c9ded5f-0ab2-455a-8dcb-855d5214af1a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.030973] env[63531]: DEBUG oslo_vmware.api [None req-932f1c32-d337-45aa-af0e-9ce7ab65093a tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Waiting for the task: (returnval){ [ 829.030973] env[63531]: value = "task-1117862" [ 829.030973] env[63531]: _type = "Task" [ 829.030973] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.043140] env[63531]: DEBUG oslo_vmware.api [None req-932f1c32-d337-45aa-af0e-9ce7ab65093a tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Task: {'id': task-1117862, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.127955] env[63531]: DEBUG nova.network.neutron [None req-3ae55c07-6554-4c57-846d-88603c4901e5 tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Updating instance_info_cache with network_info: [{"id": "b4141cf4-dab5-4033-8cc5-06966340323b", "address": "fa:16:3e:72:a7:59", "network": {"id": "7af8bbee-a43a-40b5-99c4-aee7e19b4093", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1560507257-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}, {"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.195", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "68eaa455a44b4d96936fe8273629279a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ed82fc44-f216-40a7-a68b-62d76ffdb1f8", "external-id": "nsx-vlan-transportzone-236", "segmentation_id": 236, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4141cf4-da", "ovs_interfaceid": "b4141cf4-dab5-4033-8cc5-06966340323b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.157788] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.301666] env[63531]: DEBUG nova.compute.manager [req-fb7dd3d1-fe64-46c4-b0cd-7b303cc8b23d req-e0a91e0c-530e-414d-9a0d-bf30e21ddbb4 service nova] [instance: 43c29443-1210-4ee2-95c7-1257de308287] Received event network-changed-73fd445a-3029-48b0-b3d7-7246dc21081b {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 829.302084] env[63531]: DEBUG nova.compute.manager [req-fb7dd3d1-fe64-46c4-b0cd-7b303cc8b23d req-e0a91e0c-530e-414d-9a0d-bf30e21ddbb4 service nova] [instance: 43c29443-1210-4ee2-95c7-1257de308287] Refreshing instance network info cache due to event network-changed-73fd445a-3029-48b0-b3d7-7246dc21081b. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 829.302084] env[63531]: DEBUG oslo_concurrency.lockutils [req-fb7dd3d1-fe64-46c4-b0cd-7b303cc8b23d req-e0a91e0c-530e-414d-9a0d-bf30e21ddbb4 service nova] Acquiring lock "refresh_cache-43c29443-1210-4ee2-95c7-1257de308287" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.315077] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Releasing lock "refresh_cache-43c29443-1210-4ee2-95c7-1257de308287" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.315382] env[63531]: DEBUG nova.compute.manager [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] [instance: 43c29443-1210-4ee2-95c7-1257de308287] Instance network_info: |[{"id": "73fd445a-3029-48b0-b3d7-7246dc21081b", "address": "fa:16:3e:bf:a4:b6", "network": {"id": "e01dcba8-0a6e-4458-aab6-e97c2c843bea", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-516244587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8755c58b4afa4fb08a667efc33734ce6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b7bf7d4-8e0c-4cee-84ba-244e73ef6379", "external-id": "nsx-vlan-transportzone-423", "segmentation_id": 423, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap73fd445a-30", "ovs_interfaceid": "73fd445a-3029-48b0-b3d7-7246dc21081b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 829.315673] env[63531]: DEBUG oslo_concurrency.lockutils [req-fb7dd3d1-fe64-46c4-b0cd-7b303cc8b23d req-e0a91e0c-530e-414d-9a0d-bf30e21ddbb4 service nova] Acquired lock "refresh_cache-43c29443-1210-4ee2-95c7-1257de308287" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.315853] env[63531]: DEBUG nova.network.neutron [req-fb7dd3d1-fe64-46c4-b0cd-7b303cc8b23d req-e0a91e0c-530e-414d-9a0d-bf30e21ddbb4 service nova] [instance: 43c29443-1210-4ee2-95c7-1257de308287] Refreshing network info cache for port 73fd445a-3029-48b0-b3d7-7246dc21081b {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 829.317069] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] [instance: 43c29443-1210-4ee2-95c7-1257de308287] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bf:a4:b6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3b7bf7d4-8e0c-4cee-84ba-244e73ef6379', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '73fd445a-3029-48b0-b3d7-7246dc21081b', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 829.325356] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Creating folder: Project (8755c58b4afa4fb08a667efc33734ce6). Parent ref: group-v244585. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 829.326872] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3d74309a-b145-4357-9c8f-38b1e1013763 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.338524] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Created folder: Project (8755c58b4afa4fb08a667efc33734ce6) in parent group-v244585. [ 829.338768] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Creating folder: Instances. Parent ref: group-v244690. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 829.341459] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-778964f5-3b47-4650-b6b4-5fcc70518091 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.352289] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Created folder: Instances in parent group-v244690. [ 829.352650] env[63531]: DEBUG oslo.service.loopingcall [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 829.352961] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 43c29443-1210-4ee2-95c7-1257de308287] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 829.354156] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5d2c6a6c-62be-42a1-bda6-f7de9e65d105 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.378959] env[63531]: DEBUG oslo_vmware.api [None req-13b8019c-692f-4097-a7ee-b3251d4a991c tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117861, 'name': CloneVM_Task} progress is 94%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.380365] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 829.380365] env[63531]: value = "task-1117865" [ 829.380365] env[63531]: _type = "Task" [ 829.380365] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.389959] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117865, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.463951] env[63531]: DEBUG nova.network.neutron [None req-578f74de-4af2-4ff1-9a49-a39d5b6c11ae tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Updating instance_info_cache with network_info: [{"id": "f1ef6f0e-1f62-4034-8edb-e8cc1bb98abd", "address": "fa:16:3e:3e:82:e5", "network": {"id": "dc416058-1c04-47ac-8b73-1557f7c33d1c", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1350689703-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "824ed405cc6e4868a70e82489e05e8ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "05b1253d-2b87-4158-9ff1-dafcf829f11f", "external-id": "nsx-vlan-transportzone-55", "segmentation_id": 55, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf1ef6f0e-1f", "ovs_interfaceid": "f1ef6f0e-1f62-4034-8edb-e8cc1bb98abd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.500333] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 879924d4-b465-4102-a0e3-c7b2be7ef08b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 829.545361] env[63531]: DEBUG oslo_vmware.api [None req-932f1c32-d337-45aa-af0e-9ce7ab65093a tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Task: {'id': task-1117862, 'name': PowerOffVM_Task, 'duration_secs': 0.364738} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.546444] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-932f1c32-d337-45aa-af0e-9ce7ab65093a tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 829.546444] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-932f1c32-d337-45aa-af0e-9ce7ab65093a tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 829.546569] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d817bbbb-64a7-4b17-a85e-54b71acab764 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.611327] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-932f1c32-d337-45aa-af0e-9ce7ab65093a tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 829.611681] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-932f1c32-d337-45aa-af0e-9ce7ab65093a tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 829.611971] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-932f1c32-d337-45aa-af0e-9ce7ab65093a tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Deleting the datastore file [datastore1] 5e9042b3-4dc2-4fa3-a664-c4b49a22e400 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 829.612358] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1d9c5845-26c2-4950-8079-30f9c55eca15 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.620646] env[63531]: DEBUG oslo_vmware.api [None req-932f1c32-d337-45aa-af0e-9ce7ab65093a tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Waiting for the task: (returnval){ [ 829.620646] env[63531]: value = "task-1117867" [ 829.620646] env[63531]: _type = "Task" [ 829.620646] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.629345] env[63531]: DEBUG oslo_vmware.api [None req-932f1c32-d337-45aa-af0e-9ce7ab65093a tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Task: {'id': task-1117867, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.630982] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3ae55c07-6554-4c57-846d-88603c4901e5 tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Releasing lock "refresh_cache-0197e03d-6c36-4e73-9472-8e3319eb89ce" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.631240] env[63531]: DEBUG nova.compute.manager [None req-3ae55c07-6554-4c57-846d-88603c4901e5 tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Inject network info {{(pid=63531) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7219}} [ 829.631513] env[63531]: DEBUG nova.compute.manager [None req-3ae55c07-6554-4c57-846d-88603c4901e5 tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] network_info to inject: |[{"id": "b4141cf4-dab5-4033-8cc5-06966340323b", "address": "fa:16:3e:72:a7:59", "network": {"id": "7af8bbee-a43a-40b5-99c4-aee7e19b4093", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1560507257-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}, {"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.195", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "68eaa455a44b4d96936fe8273629279a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ed82fc44-f216-40a7-a68b-62d76ffdb1f8", "external-id": "nsx-vlan-transportzone-236", "segmentation_id": 236, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4141cf4-da", "ovs_interfaceid": "b4141cf4-dab5-4033-8cc5-06966340323b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7220}} [ 829.636219] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-3ae55c07-6554-4c57-846d-88603c4901e5 tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Reconfiguring VM instance to set the machine id {{(pid=63531) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1802}} [ 829.636529] env[63531]: DEBUG oslo_concurrency.lockutils [req-ea2a35b8-a2da-44b0-8942-9a1e5b913e35 req-6202a6df-9f65-47b7-87db-2e49af35b64b service nova] Acquired lock "refresh_cache-0197e03d-6c36-4e73-9472-8e3319eb89ce" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.636711] env[63531]: DEBUG nova.network.neutron [req-ea2a35b8-a2da-44b0-8942-9a1e5b913e35 req-6202a6df-9f65-47b7-87db-2e49af35b64b service nova] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Refreshing network info cache for port b4141cf4-dab5-4033-8cc5-06966340323b {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 829.637974] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2e66b6a8-6809-4d0a-a0e3-fa86e491028d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.651297] env[63531]: DEBUG nova.objects.instance [None req-c7269493-588f-4a5a-94b1-f08c0a54a60a tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Lazy-loading 'flavor' on Instance uuid 0197e03d-6c36-4e73-9472-8e3319eb89ce {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 829.658171] env[63531]: DEBUG oslo_vmware.api [None req-3ae55c07-6554-4c57-846d-88603c4901e5 tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Waiting for the task: (returnval){ [ 829.658171] env[63531]: value = "task-1117868" [ 829.658171] env[63531]: _type = "Task" [ 829.658171] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.667982] env[63531]: DEBUG oslo_vmware.api [None req-3ae55c07-6554-4c57-846d-88603c4901e5 tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Task: {'id': task-1117868, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.882901] env[63531]: DEBUG oslo_vmware.api [None req-13b8019c-692f-4097-a7ee-b3251d4a991c tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117861, 'name': CloneVM_Task} progress is 94%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.890787] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117865, 'name': CreateVM_Task, 'duration_secs': 0.402736} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.890955] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 43c29443-1210-4ee2-95c7-1257de308287] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 829.892915] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.893103] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.893435] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 829.893700] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ad7dc67f-dbaa-471b-8314-c584321c5639 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.899103] env[63531]: DEBUG oslo_vmware.api [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Waiting for the task: (returnval){ [ 829.899103] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]520ac331-7734-67a9-a254-be02da870402" [ 829.899103] env[63531]: _type = "Task" [ 829.899103] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.907469] env[63531]: DEBUG oslo_vmware.api [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]520ac331-7734-67a9-a254-be02da870402, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.967065] env[63531]: DEBUG oslo_concurrency.lockutils [None req-578f74de-4af2-4ff1-9a49-a39d5b6c11ae tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Releasing lock "refresh_cache-e0074e2e-e8bd-4bcb-9ad9-271a10e079e4" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.004326] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance e95d3273-8216-47cc-95b6-99301366a827 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 830.004536] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Total usable vcpus: 48, total allocated vcpus: 16 {{(pid=63531) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 830.004684] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3584MB phys_disk=200GB used_disk=16GB total_vcpus=48 used_vcpus=16 pci_stats=[] {{(pid=63531) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 830.132437] env[63531]: DEBUG oslo_vmware.api [None req-932f1c32-d337-45aa-af0e-9ce7ab65093a tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Task: {'id': task-1117867, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.177226} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.132714] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-932f1c32-d337-45aa-af0e-9ce7ab65093a tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 830.133769] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-932f1c32-d337-45aa-af0e-9ce7ab65093a tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 830.133769] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-932f1c32-d337-45aa-af0e-9ce7ab65093a tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 830.133769] env[63531]: INFO nova.compute.manager [None req-932f1c32-d337-45aa-af0e-9ce7ab65093a tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Took 1.12 seconds to destroy the instance on the hypervisor. [ 830.133769] env[63531]: DEBUG oslo.service.loopingcall [None req-932f1c32-d337-45aa-af0e-9ce7ab65093a tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 830.133769] env[63531]: DEBUG nova.compute.manager [-] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 830.133769] env[63531]: DEBUG nova.network.neutron [-] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 830.168153] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c7269493-588f-4a5a-94b1-f08c0a54a60a tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Acquiring lock "refresh_cache-0197e03d-6c36-4e73-9472-8e3319eb89ce" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.174283] env[63531]: DEBUG oslo_vmware.api [None req-3ae55c07-6554-4c57-846d-88603c4901e5 tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Task: {'id': task-1117868, 'name': ReconfigVM_Task, 'duration_secs': 0.174667} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.176875] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-3ae55c07-6554-4c57-846d-88603c4901e5 tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Reconfigured VM instance to set the machine id {{(pid=63531) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1805}} [ 830.250240] env[63531]: DEBUG nova.network.neutron [req-fb7dd3d1-fe64-46c4-b0cd-7b303cc8b23d req-e0a91e0c-530e-414d-9a0d-bf30e21ddbb4 service nova] [instance: 43c29443-1210-4ee2-95c7-1257de308287] Updated VIF entry in instance network info cache for port 73fd445a-3029-48b0-b3d7-7246dc21081b. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 830.250549] env[63531]: DEBUG nova.network.neutron [req-fb7dd3d1-fe64-46c4-b0cd-7b303cc8b23d req-e0a91e0c-530e-414d-9a0d-bf30e21ddbb4 service nova] [instance: 43c29443-1210-4ee2-95c7-1257de308287] Updating instance_info_cache with network_info: [{"id": "73fd445a-3029-48b0-b3d7-7246dc21081b", "address": "fa:16:3e:bf:a4:b6", "network": {"id": "e01dcba8-0a6e-4458-aab6-e97c2c843bea", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-516244587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8755c58b4afa4fb08a667efc33734ce6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b7bf7d4-8e0c-4cee-84ba-244e73ef6379", "external-id": "nsx-vlan-transportzone-423", "segmentation_id": 423, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap73fd445a-30", "ovs_interfaceid": "73fd445a-3029-48b0-b3d7-7246dc21081b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.384088] env[63531]: DEBUG oslo_vmware.api [None req-13b8019c-692f-4097-a7ee-b3251d4a991c tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117861, 'name': CloneVM_Task, 'duration_secs': 1.415949} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.385087] env[63531]: INFO nova.virt.vmwareapi.vmops [None req-13b8019c-692f-4097-a7ee-b3251d4a991c tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Created linked-clone VM from snapshot [ 830.385274] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-655a6fd9-14ec-4220-98e8-80040a8c4b7f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.397429] env[63531]: DEBUG nova.virt.vmwareapi.images [None req-13b8019c-692f-4097-a7ee-b3251d4a991c tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Uploading image 2ef8015f-f209-49ec-9140-1466c4081402 {{(pid=63531) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 830.410250] env[63531]: DEBUG oslo_vmware.api [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]520ac331-7734-67a9-a254-be02da870402, 'name': SearchDatastore_Task, 'duration_secs': 0.012341} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.412593] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.412823] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] [instance: 43c29443-1210-4ee2-95c7-1257de308287] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 830.413104] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.413262] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.413452] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 830.414282] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07eea4ef-3124-4bf1-a843-e8dcc8a92d55 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.419330] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a09eb369-9402-44e3-984f-e4ec657f5599 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.426676] env[63531]: DEBUG oslo_vmware.rw_handles [None req-13b8019c-692f-4097-a7ee-b3251d4a991c tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 830.426676] env[63531]: value = "vm-244689" [ 830.426676] env[63531]: _type = "VirtualMachine" [ 830.426676] env[63531]: }. {{(pid=63531) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 830.427663] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e64f708-3536-4f0f-9431-367d101b15d7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.430678] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-41b18d07-c2f6-4cb3-b1b7-4c204beb8aa6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.432968] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 830.433169] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 830.434373] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-506c07f4-ec43-4ed2-a9ec-b6ac6b43cb43 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.466013] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b28a79f-21de-48d8-87b2-4ef22c95e48b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.468471] env[63531]: DEBUG oslo_vmware.rw_handles [None req-13b8019c-692f-4097-a7ee-b3251d4a991c tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Lease: (returnval){ [ 830.468471] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52964fe0-164f-8e05-c8bf-9a17ade5fdb7" [ 830.468471] env[63531]: _type = "HttpNfcLease" [ 830.468471] env[63531]: } obtained for exporting VM: (result){ [ 830.468471] env[63531]: value = "vm-244689" [ 830.468471] env[63531]: _type = "VirtualMachine" [ 830.468471] env[63531]: }. {{(pid=63531) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 830.468782] env[63531]: DEBUG oslo_vmware.api [None req-13b8019c-692f-4097-a7ee-b3251d4a991c tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Waiting for the lease: (returnval){ [ 830.468782] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52964fe0-164f-8e05-c8bf-9a17ade5fdb7" [ 830.468782] env[63531]: _type = "HttpNfcLease" [ 830.468782] env[63531]: } to be ready. {{(pid=63531) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 830.468941] env[63531]: DEBUG oslo_vmware.api [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Waiting for the task: (returnval){ [ 830.468941] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5284b37f-a740-9b65-c43d-9501c1fdaedf" [ 830.468941] env[63531]: _type = "Task" [ 830.468941] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.478976] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-068e12bb-146a-4ac1-b30b-88b283ec5241 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.483927] env[63531]: DEBUG nova.compute.manager [req-18fea2cd-5743-4d46-aa9e-de3d3fba617a req-2f0902c3-f47b-471e-8b98-cd8178904f31 service nova] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Received event network-vif-deleted-f178c862-e753-40d4-b7cd-3eabe7eda39e {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 830.484136] env[63531]: INFO nova.compute.manager [req-18fea2cd-5743-4d46-aa9e-de3d3fba617a req-2f0902c3-f47b-471e-8b98-cd8178904f31 service nova] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Neutron deleted interface f178c862-e753-40d4-b7cd-3eabe7eda39e; detaching it from the instance and deleting it from the info cache [ 830.484313] env[63531]: DEBUG nova.network.neutron [req-18fea2cd-5743-4d46-aa9e-de3d3fba617a req-2f0902c3-f47b-471e-8b98-cd8178904f31 service nova] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.494039] env[63531]: DEBUG oslo_vmware.api [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5284b37f-a740-9b65-c43d-9501c1fdaedf, 'name': SearchDatastore_Task, 'duration_secs': 0.025813} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.494334] env[63531]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 830.494334] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52964fe0-164f-8e05-c8bf-9a17ade5fdb7" [ 830.494334] env[63531]: _type = "HttpNfcLease" [ 830.494334] env[63531]: } is ready. {{(pid=63531) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 830.496058] env[63531]: DEBUG oslo_vmware.rw_handles [None req-13b8019c-692f-4097-a7ee-b3251d4a991c tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 830.496058] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52964fe0-164f-8e05-c8bf-9a17ade5fdb7" [ 830.496058] env[63531]: _type = "HttpNfcLease" [ 830.496058] env[63531]: }. {{(pid=63531) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 830.496554] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-088d8f75-eaa0-484a-93ae-1a8210820ecc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.501374] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69e52e28-ec3b-437c-9a20-e52ec5a8f2ec {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.505880] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-578f74de-4af2-4ff1-9a49-a39d5b6c11ae tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 830.515712] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e0344dea-abd6-4f43-937b-db22b2e36e9e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.517303] env[63531]: DEBUG nova.compute.provider_tree [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 830.521478] env[63531]: DEBUG oslo_vmware.api [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Waiting for the task: (returnval){ [ 830.521478] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52e46af4-df12-d796-481e-7a28d4218d9b" [ 830.521478] env[63531]: _type = "Task" [ 830.521478] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.526017] env[63531]: DEBUG oslo_vmware.rw_handles [None req-13b8019c-692f-4097-a7ee-b3251d4a991c tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524335f2-851d-a3af-b6a7-97a78a8f2d6e/disk-0.vmdk from lease info. {{(pid=63531) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 830.526017] env[63531]: DEBUG oslo_vmware.rw_handles [None req-13b8019c-692f-4097-a7ee-b3251d4a991c tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524335f2-851d-a3af-b6a7-97a78a8f2d6e/disk-0.vmdk for reading. {{(pid=63531) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 830.529755] env[63531]: DEBUG oslo_vmware.api [None req-578f74de-4af2-4ff1-9a49-a39d5b6c11ae tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Waiting for the task: (returnval){ [ 830.529755] env[63531]: value = "task-1117870" [ 830.529755] env[63531]: _type = "Task" [ 830.529755] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.595692] env[63531]: DEBUG nova.network.neutron [req-ea2a35b8-a2da-44b0-8942-9a1e5b913e35 req-6202a6df-9f65-47b7-87db-2e49af35b64b service nova] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Updated VIF entry in instance network info cache for port b4141cf4-dab5-4033-8cc5-06966340323b. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 830.596097] env[63531]: DEBUG nova.network.neutron [req-ea2a35b8-a2da-44b0-8942-9a1e5b913e35 req-6202a6df-9f65-47b7-87db-2e49af35b64b service nova] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Updating instance_info_cache with network_info: [{"id": "b4141cf4-dab5-4033-8cc5-06966340323b", "address": "fa:16:3e:72:a7:59", "network": {"id": "7af8bbee-a43a-40b5-99c4-aee7e19b4093", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1560507257-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}, {"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.195", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "68eaa455a44b4d96936fe8273629279a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ed82fc44-f216-40a7-a68b-62d76ffdb1f8", "external-id": "nsx-vlan-transportzone-236", "segmentation_id": 236, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4141cf4-da", "ovs_interfaceid": "b4141cf4-dab5-4033-8cc5-06966340323b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.601050] env[63531]: DEBUG oslo_vmware.api [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52e46af4-df12-d796-481e-7a28d4218d9b, 'name': SearchDatastore_Task, 'duration_secs': 0.017546} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.602280] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.602617] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] 43c29443-1210-4ee2-95c7-1257de308287/43c29443-1210-4ee2-95c7-1257de308287.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 830.602942] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f66391cd-12dd-4b1b-a66e-225795d05a78 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.610998] env[63531]: DEBUG oslo_vmware.api [None req-578f74de-4af2-4ff1-9a49-a39d5b6c11ae tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117870, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.615806] env[63531]: DEBUG oslo_vmware.api [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Waiting for the task: (returnval){ [ 830.615806] env[63531]: value = "task-1117871" [ 830.615806] env[63531]: _type = "Task" [ 830.615806] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.627482] env[63531]: DEBUG oslo_vmware.api [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Task: {'id': task-1117871, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.645399] env[63531]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-58caed76-487c-4a29-942b-696a09e9ecb2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.752702] env[63531]: DEBUG oslo_concurrency.lockutils [req-fb7dd3d1-fe64-46c4-b0cd-7b303cc8b23d req-e0a91e0c-530e-414d-9a0d-bf30e21ddbb4 service nova] Releasing lock "refresh_cache-43c29443-1210-4ee2-95c7-1257de308287" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.919482] env[63531]: DEBUG nova.network.neutron [-] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.995972] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1caba3b4-4473-4a9e-a818-509cd3fe1c79 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.010626] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-448c59a3-bed6-4303-a818-2966dea7ba47 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.030352] env[63531]: DEBUG nova.scheduler.client.report [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 831.049152] env[63531]: DEBUG oslo_vmware.api [None req-578f74de-4af2-4ff1-9a49-a39d5b6c11ae tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117870, 'name': PowerOffVM_Task, 'duration_secs': 0.187502} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.066667] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-578f74de-4af2-4ff1-9a49-a39d5b6c11ae tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 831.068380] env[63531]: DEBUG nova.compute.manager [req-18fea2cd-5743-4d46-aa9e-de3d3fba617a req-2f0902c3-f47b-471e-8b98-cd8178904f31 service nova] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Detach interface failed, port_id=f178c862-e753-40d4-b7cd-3eabe7eda39e, reason: Instance 5e9042b3-4dc2-4fa3-a664-c4b49a22e400 could not be found. {{(pid=63531) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 831.069630] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e2b864e-26c3-4c1f-b90a-383736556394 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.092868] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c06200ec-d90c-477f-9240-0e695b8f6c5c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.103340] env[63531]: DEBUG oslo_concurrency.lockutils [req-ea2a35b8-a2da-44b0-8942-9a1e5b913e35 req-6202a6df-9f65-47b7-87db-2e49af35b64b service nova] Releasing lock "refresh_cache-0197e03d-6c36-4e73-9472-8e3319eb89ce" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.104014] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c7269493-588f-4a5a-94b1-f08c0a54a60a tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Acquired lock "refresh_cache-0197e03d-6c36-4e73-9472-8e3319eb89ce" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.124877] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-578f74de-4af2-4ff1-9a49-a39d5b6c11ae tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 831.125301] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9ae1ea1a-1eb3-4de5-8b1f-d15e770c34cc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.130411] env[63531]: DEBUG oslo_vmware.api [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Task: {'id': task-1117871, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.49844} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.131394] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] 43c29443-1210-4ee2-95c7-1257de308287/43c29443-1210-4ee2-95c7-1257de308287.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 831.131769] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] [instance: 43c29443-1210-4ee2-95c7-1257de308287] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 831.132077] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fd647bed-311f-4511-a59c-16e453ef4b9e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.135646] env[63531]: DEBUG oslo_vmware.api [None req-578f74de-4af2-4ff1-9a49-a39d5b6c11ae tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Waiting for the task: (returnval){ [ 831.135646] env[63531]: value = "task-1117872" [ 831.135646] env[63531]: _type = "Task" [ 831.135646] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.140615] env[63531]: DEBUG oslo_vmware.api [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Waiting for the task: (returnval){ [ 831.140615] env[63531]: value = "task-1117873" [ 831.140615] env[63531]: _type = "Task" [ 831.140615] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.148379] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-578f74de-4af2-4ff1-9a49-a39d5b6c11ae tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] VM already powered off {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 831.148725] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-578f74de-4af2-4ff1-9a49-a39d5b6c11ae tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 831.149081] env[63531]: DEBUG oslo_concurrency.lockutils [None req-578f74de-4af2-4ff1-9a49-a39d5b6c11ae tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.149301] env[63531]: DEBUG oslo_concurrency.lockutils [None req-578f74de-4af2-4ff1-9a49-a39d5b6c11ae tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.149543] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-578f74de-4af2-4ff1-9a49-a39d5b6c11ae tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 831.152977] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bd59dec1-e1d6-429f-a7b0-2c94c8dcc20f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.155508] env[63531]: DEBUG oslo_vmware.api [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Task: {'id': task-1117873, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.163277] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-578f74de-4af2-4ff1-9a49-a39d5b6c11ae tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 831.163524] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-578f74de-4af2-4ff1-9a49-a39d5b6c11ae tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 831.164465] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ce606750-6ad8-4241-9959-619d64e7601f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.172622] env[63531]: DEBUG oslo_vmware.api [None req-578f74de-4af2-4ff1-9a49-a39d5b6c11ae tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Waiting for the task: (returnval){ [ 831.172622] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52a2d444-2c4e-8571-9db8-25abc92eadfc" [ 831.172622] env[63531]: _type = "Task" [ 831.172622] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.181587] env[63531]: DEBUG oslo_vmware.api [None req-578f74de-4af2-4ff1-9a49-a39d5b6c11ae tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52a2d444-2c4e-8571-9db8-25abc92eadfc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.348056] env[63531]: DEBUG nova.compute.manager [None req-92007411-3bac-410d-a9c6-c0b9a37679fa tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 831.349429] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-348db648-b10d-4d8f-801c-3f86b72af6e6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.422583] env[63531]: INFO nova.compute.manager [-] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Took 1.29 seconds to deallocate network for instance. [ 831.538910] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63531) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 831.539437] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 7.651s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.539767] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 38.161s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.542330] env[63531]: INFO nova.compute.claims [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 831.549695] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 831.549910] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Cleaning up deleted instances {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 831.653742] env[63531]: DEBUG oslo_vmware.api [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Task: {'id': task-1117873, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073405} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.653742] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] [instance: 43c29443-1210-4ee2-95c7-1257de308287] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 831.653742] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bb06a3e-10b1-4122-a9b7-e78532ee3dc1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.676621] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] [instance: 43c29443-1210-4ee2-95c7-1257de308287] Reconfiguring VM instance instance-0000003e to attach disk [datastore2] 43c29443-1210-4ee2-95c7-1257de308287/43c29443-1210-4ee2-95c7-1257de308287.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 831.677167] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-380100c0-b039-469d-9cf9-63d72a54b174 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.704185] env[63531]: DEBUG oslo_vmware.api [None req-578f74de-4af2-4ff1-9a49-a39d5b6c11ae tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52a2d444-2c4e-8571-9db8-25abc92eadfc, 'name': SearchDatastore_Task, 'duration_secs': 0.009564} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.709022] env[63531]: DEBUG oslo_vmware.api [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Waiting for the task: (returnval){ [ 831.709022] env[63531]: value = "task-1117874" [ 831.709022] env[63531]: _type = "Task" [ 831.709022] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.709022] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-21f1e0ac-3c99-46ca-9984-d83749c16371 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.714565] env[63531]: DEBUG oslo_vmware.api [None req-578f74de-4af2-4ff1-9a49-a39d5b6c11ae tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Waiting for the task: (returnval){ [ 831.714565] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52034d08-0a27-f1c4-15a5-1fbc74a1fd82" [ 831.714565] env[63531]: _type = "Task" [ 831.714565] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.725421] env[63531]: DEBUG oslo_vmware.api [None req-578f74de-4af2-4ff1-9a49-a39d5b6c11ae tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52034d08-0a27-f1c4-15a5-1fbc74a1fd82, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.860556] env[63531]: INFO nova.compute.manager [None req-92007411-3bac-410d-a9c6-c0b9a37679fa tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] instance snapshotting [ 831.861717] env[63531]: DEBUG nova.objects.instance [None req-92007411-3bac-410d-a9c6-c0b9a37679fa tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lazy-loading 'flavor' on Instance uuid 9a3fad46-a15a-451c-bdab-a3c8cc8add07 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 831.935738] env[63531]: DEBUG oslo_concurrency.lockutils [None req-932f1c32-d337-45aa-af0e-9ce7ab65093a tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.015294] env[63531]: DEBUG nova.network.neutron [None req-c7269493-588f-4a5a-94b1-f08c0a54a60a tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 832.040161] env[63531]: DEBUG nova.compute.manager [req-5280cb6a-73a2-4b6f-86bf-aec5c72333e7 req-0ef86b50-799c-417b-8adb-20f3e060db04 service nova] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Received event network-changed-b4141cf4-dab5-4033-8cc5-06966340323b {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 832.040555] env[63531]: DEBUG nova.compute.manager [req-5280cb6a-73a2-4b6f-86bf-aec5c72333e7 req-0ef86b50-799c-417b-8adb-20f3e060db04 service nova] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Refreshing instance network info cache due to event network-changed-b4141cf4-dab5-4033-8cc5-06966340323b. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 832.040870] env[63531]: DEBUG oslo_concurrency.lockutils [req-5280cb6a-73a2-4b6f-86bf-aec5c72333e7 req-0ef86b50-799c-417b-8adb-20f3e060db04 service nova] Acquiring lock "refresh_cache-0197e03d-6c36-4e73-9472-8e3319eb89ce" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.057394] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] There are 15 instances to clean {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 832.057670] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: dccb2335-4220-4570-861b-16ff2d8f5a85] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 832.219663] env[63531]: DEBUG oslo_vmware.api [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Task: {'id': task-1117874, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.228453] env[63531]: DEBUG oslo_vmware.api [None req-578f74de-4af2-4ff1-9a49-a39d5b6c11ae tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52034d08-0a27-f1c4-15a5-1fbc74a1fd82, 'name': SearchDatastore_Task, 'duration_secs': 0.026908} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.229093] env[63531]: DEBUG oslo_concurrency.lockutils [None req-578f74de-4af2-4ff1-9a49-a39d5b6c11ae tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.229512] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-578f74de-4af2-4ff1-9a49-a39d5b6c11ae tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] e0074e2e-e8bd-4bcb-9ad9-271a10e079e4/9d550399-588c-4bd5-8436-eba9b523bd13-rescue.vmdk. {{(pid=63531) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 832.230653] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4f23850c-6d3c-4664-90d9-a6ad8ea08977 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.236898] env[63531]: DEBUG oslo_vmware.api [None req-578f74de-4af2-4ff1-9a49-a39d5b6c11ae tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Waiting for the task: (returnval){ [ 832.236898] env[63531]: value = "task-1117875" [ 832.236898] env[63531]: _type = "Task" [ 832.236898] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.244683] env[63531]: DEBUG oslo_vmware.api [None req-578f74de-4af2-4ff1-9a49-a39d5b6c11ae tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117875, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.369587] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4370cb67-3cc9-4257-a99f-cd7f7a6b274b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.390163] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-728f0eaf-c7dd-4c11-a789-2663baef374c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.561924] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: e1693594-d8bb-4002-b71c-7013efb81d16] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 832.720014] env[63531]: DEBUG oslo_vmware.api [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Task: {'id': task-1117874, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.749365] env[63531]: DEBUG oslo_vmware.api [None req-578f74de-4af2-4ff1-9a49-a39d5b6c11ae tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117875, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.454996} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.751935] env[63531]: INFO nova.virt.vmwareapi.ds_util [None req-578f74de-4af2-4ff1-9a49-a39d5b6c11ae tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] e0074e2e-e8bd-4bcb-9ad9-271a10e079e4/9d550399-588c-4bd5-8436-eba9b523bd13-rescue.vmdk. [ 832.752928] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7076ea8a-7cbd-42ac-9249-37ec84e877bf {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.777960] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-578f74de-4af2-4ff1-9a49-a39d5b6c11ae tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Reconfiguring VM instance instance-0000003d to attach disk [datastore2] e0074e2e-e8bd-4bcb-9ad9-271a10e079e4/9d550399-588c-4bd5-8436-eba9b523bd13-rescue.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 832.780851] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1ba5f552-b526-47c5-8965-db54f86effc0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.799218] env[63531]: DEBUG oslo_vmware.api [None req-578f74de-4af2-4ff1-9a49-a39d5b6c11ae tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Waiting for the task: (returnval){ [ 832.799218] env[63531]: value = "task-1117876" [ 832.799218] env[63531]: _type = "Task" [ 832.799218] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.812313] env[63531]: DEBUG oslo_vmware.api [None req-578f74de-4af2-4ff1-9a49-a39d5b6c11ae tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117876, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.875699] env[63531]: DEBUG nova.network.neutron [None req-c7269493-588f-4a5a-94b1-f08c0a54a60a tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Updating instance_info_cache with network_info: [{"id": "b4141cf4-dab5-4033-8cc5-06966340323b", "address": "fa:16:3e:72:a7:59", "network": {"id": "7af8bbee-a43a-40b5-99c4-aee7e19b4093", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1560507257-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.195", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "68eaa455a44b4d96936fe8273629279a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ed82fc44-f216-40a7-a68b-62d76ffdb1f8", "external-id": "nsx-vlan-transportzone-236", "segmentation_id": 236, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4141cf4-da", "ovs_interfaceid": "b4141cf4-dab5-4033-8cc5-06966340323b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 832.904128] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-92007411-3bac-410d-a9c6-c0b9a37679fa tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Creating Snapshot of the VM instance {{(pid=63531) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 832.904128] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-c31e8187-21b6-4f5f-a61f-26de1c49302e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.912802] env[63531]: DEBUG oslo_vmware.api [None req-92007411-3bac-410d-a9c6-c0b9a37679fa tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 832.912802] env[63531]: value = "task-1117877" [ 832.912802] env[63531]: _type = "Task" [ 832.912802] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.921861] env[63531]: DEBUG oslo_vmware.api [None req-92007411-3bac-410d-a9c6-c0b9a37679fa tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1117877, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.002722] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2162057-00f9-4a46-a1bb-0f3df34c66a8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.011450] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-501da32d-254c-4dfc-acfe-4a450b8efe02 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.043850] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d7225dc-2f4c-4864-bc7e-839dc6e2125e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.054572] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27170bdf-eda8-4b71-8d39-243d382d5af2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.069890] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: e4f2ab48-8fbd-4de8-bed9-7507e4cd1b74] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 833.072402] env[63531]: DEBUG nova.compute.provider_tree [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 833.221056] env[63531]: DEBUG oslo_vmware.api [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Task: {'id': task-1117874, 'name': ReconfigVM_Task, 'duration_secs': 1.317799} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.221559] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] [instance: 43c29443-1210-4ee2-95c7-1257de308287] Reconfigured VM instance instance-0000003e to attach disk [datastore2] 43c29443-1210-4ee2-95c7-1257de308287/43c29443-1210-4ee2-95c7-1257de308287.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 833.222101] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bb33ba61-b0a7-4eed-aad5-1006b5e2d98b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.228696] env[63531]: DEBUG oslo_vmware.api [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Waiting for the task: (returnval){ [ 833.228696] env[63531]: value = "task-1117878" [ 833.228696] env[63531]: _type = "Task" [ 833.228696] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.237274] env[63531]: DEBUG oslo_vmware.api [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Task: {'id': task-1117878, 'name': Rename_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.311437] env[63531]: DEBUG oslo_vmware.api [None req-578f74de-4af2-4ff1-9a49-a39d5b6c11ae tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117876, 'name': ReconfigVM_Task, 'duration_secs': 0.290364} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.311807] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-578f74de-4af2-4ff1-9a49-a39d5b6c11ae tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Reconfigured VM instance instance-0000003d to attach disk [datastore2] e0074e2e-e8bd-4bcb-9ad9-271a10e079e4/9d550399-588c-4bd5-8436-eba9b523bd13-rescue.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 833.312783] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c5bd6b4-e819-4893-b8ab-484c7f6e0e9b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.339625] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f333162a-066e-4f4a-aed7-802b93aa89fe {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.359501] env[63531]: DEBUG oslo_vmware.api [None req-578f74de-4af2-4ff1-9a49-a39d5b6c11ae tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Waiting for the task: (returnval){ [ 833.359501] env[63531]: value = "task-1117879" [ 833.359501] env[63531]: _type = "Task" [ 833.359501] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.368246] env[63531]: DEBUG oslo_vmware.api [None req-578f74de-4af2-4ff1-9a49-a39d5b6c11ae tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117879, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.379679] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c7269493-588f-4a5a-94b1-f08c0a54a60a tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Releasing lock "refresh_cache-0197e03d-6c36-4e73-9472-8e3319eb89ce" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.380053] env[63531]: DEBUG nova.compute.manager [None req-c7269493-588f-4a5a-94b1-f08c0a54a60a tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Inject network info {{(pid=63531) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7219}} [ 833.380417] env[63531]: DEBUG nova.compute.manager [None req-c7269493-588f-4a5a-94b1-f08c0a54a60a tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] network_info to inject: |[{"id": "b4141cf4-dab5-4033-8cc5-06966340323b", "address": "fa:16:3e:72:a7:59", "network": {"id": "7af8bbee-a43a-40b5-99c4-aee7e19b4093", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1560507257-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.195", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "68eaa455a44b4d96936fe8273629279a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ed82fc44-f216-40a7-a68b-62d76ffdb1f8", "external-id": "nsx-vlan-transportzone-236", "segmentation_id": 236, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4141cf4-da", "ovs_interfaceid": "b4141cf4-dab5-4033-8cc5-06966340323b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7220}} [ 833.386057] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c7269493-588f-4a5a-94b1-f08c0a54a60a tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Reconfiguring VM instance to set the machine id {{(pid=63531) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1802}} [ 833.386372] env[63531]: DEBUG oslo_concurrency.lockutils [req-5280cb6a-73a2-4b6f-86bf-aec5c72333e7 req-0ef86b50-799c-417b-8adb-20f3e060db04 service nova] Acquired lock "refresh_cache-0197e03d-6c36-4e73-9472-8e3319eb89ce" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.386578] env[63531]: DEBUG nova.network.neutron [req-5280cb6a-73a2-4b6f-86bf-aec5c72333e7 req-0ef86b50-799c-417b-8adb-20f3e060db04 service nova] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Refreshing network info cache for port b4141cf4-dab5-4033-8cc5-06966340323b {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 833.387870] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-297a0588-e7d9-405c-8cc3-76cb33fc7993 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.403199] env[63531]: DEBUG oslo_vmware.api [None req-c7269493-588f-4a5a-94b1-f08c0a54a60a tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Waiting for the task: (returnval){ [ 833.403199] env[63531]: value = "task-1117880" [ 833.403199] env[63531]: _type = "Task" [ 833.403199] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.411288] env[63531]: DEBUG oslo_vmware.api [None req-c7269493-588f-4a5a-94b1-f08c0a54a60a tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Task: {'id': task-1117880, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.421680] env[63531]: DEBUG oslo_vmware.api [None req-92007411-3bac-410d-a9c6-c0b9a37679fa tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1117877, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.550287] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ead7921a-cb8c-490f-a3ab-414729b031bc tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Acquiring lock "0197e03d-6c36-4e73-9472-8e3319eb89ce" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.550599] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ead7921a-cb8c-490f-a3ab-414729b031bc tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Lock "0197e03d-6c36-4e73-9472-8e3319eb89ce" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.550823] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ead7921a-cb8c-490f-a3ab-414729b031bc tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Acquiring lock "0197e03d-6c36-4e73-9472-8e3319eb89ce-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.551021] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ead7921a-cb8c-490f-a3ab-414729b031bc tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Lock "0197e03d-6c36-4e73-9472-8e3319eb89ce-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.551202] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ead7921a-cb8c-490f-a3ab-414729b031bc tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Lock "0197e03d-6c36-4e73-9472-8e3319eb89ce-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.553538] env[63531]: INFO nova.compute.manager [None req-ead7921a-cb8c-490f-a3ab-414729b031bc tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Terminating instance [ 833.555449] env[63531]: DEBUG nova.compute.manager [None req-ead7921a-cb8c-490f-a3ab-414729b031bc tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 833.555647] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-ead7921a-cb8c-490f-a3ab-414729b031bc tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 833.556501] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cb23738-a610-442b-a0f4-71a1d0a5d8fb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.566236] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-ead7921a-cb8c-490f-a3ab-414729b031bc tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 833.566554] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bab98374-298c-4a0d-89ca-24f796aee9c0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.573800] env[63531]: DEBUG oslo_vmware.api [None req-ead7921a-cb8c-490f-a3ab-414729b031bc tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Waiting for the task: (returnval){ [ 833.573800] env[63531]: value = "task-1117881" [ 833.573800] env[63531]: _type = "Task" [ 833.573800] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.574237] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 0af99240-fa7a-4eed-a729-a5ae98c41cf9] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 833.576607] env[63531]: DEBUG nova.scheduler.client.report [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 833.588169] env[63531]: DEBUG oslo_vmware.api [None req-ead7921a-cb8c-490f-a3ab-414729b031bc tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Task: {'id': task-1117881, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.739284] env[63531]: DEBUG oslo_vmware.api [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Task: {'id': task-1117878, 'name': Rename_Task, 'duration_secs': 0.155598} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.739768] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] [instance: 43c29443-1210-4ee2-95c7-1257de308287] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 833.740173] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e1370e9f-0f21-41f8-ba62-703196dabaad {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.746738] env[63531]: DEBUG oslo_vmware.api [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Waiting for the task: (returnval){ [ 833.746738] env[63531]: value = "task-1117882" [ 833.746738] env[63531]: _type = "Task" [ 833.746738] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.755648] env[63531]: DEBUG oslo_vmware.api [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Task: {'id': task-1117882, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.872657] env[63531]: DEBUG oslo_vmware.api [None req-578f74de-4af2-4ff1-9a49-a39d5b6c11ae tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117879, 'name': ReconfigVM_Task, 'duration_secs': 0.150483} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.873144] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-578f74de-4af2-4ff1-9a49-a39d5b6c11ae tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 833.873531] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-114279ce-efcc-4cbb-876b-4270b427c1a8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.881120] env[63531]: DEBUG oslo_vmware.api [None req-578f74de-4af2-4ff1-9a49-a39d5b6c11ae tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Waiting for the task: (returnval){ [ 833.881120] env[63531]: value = "task-1117883" [ 833.881120] env[63531]: _type = "Task" [ 833.881120] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.892524] env[63531]: DEBUG oslo_vmware.api [None req-578f74de-4af2-4ff1-9a49-a39d5b6c11ae tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117883, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.915850] env[63531]: DEBUG oslo_vmware.api [None req-c7269493-588f-4a5a-94b1-f08c0a54a60a tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Task: {'id': task-1117880, 'name': ReconfigVM_Task, 'duration_secs': 0.182086} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.921165] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c7269493-588f-4a5a-94b1-f08c0a54a60a tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Reconfigured VM instance to set the machine id {{(pid=63531) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1805}} [ 833.935613] env[63531]: DEBUG oslo_vmware.api [None req-92007411-3bac-410d-a9c6-c0b9a37679fa tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1117877, 'name': CreateSnapshot_Task, 'duration_secs': 0.66719} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.935613] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-92007411-3bac-410d-a9c6-c0b9a37679fa tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Created Snapshot of the VM instance {{(pid=63531) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 833.935613] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-425665ec-38d5-483b-8ae3-25257e1d065c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.083272] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 410ddf7a-e45b-4769-a3db-5363ce2096a7] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 834.085674] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.546s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.086273] env[63531]: DEBUG nova.compute.manager [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 834.090013] env[63531]: DEBUG oslo_concurrency.lockutils [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 37.134s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.091724] env[63531]: INFO nova.compute.claims [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] [instance: 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 834.099063] env[63531]: DEBUG oslo_vmware.api [None req-ead7921a-cb8c-490f-a3ab-414729b031bc tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Task: {'id': task-1117881, 'name': PowerOffVM_Task, 'duration_secs': 0.224735} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.099715] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-ead7921a-cb8c-490f-a3ab-414729b031bc tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 834.099944] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-ead7921a-cb8c-490f-a3ab-414729b031bc tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 834.100325] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-32798814-dc75-4d30-b269-c7fff56d8ca9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.174627] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-ead7921a-cb8c-490f-a3ab-414729b031bc tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 834.174865] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-ead7921a-cb8c-490f-a3ab-414729b031bc tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 834.175077] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-ead7921a-cb8c-490f-a3ab-414729b031bc tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Deleting the datastore file [datastore1] 0197e03d-6c36-4e73-9472-8e3319eb89ce {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 834.175392] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dd5f8a3d-aa09-4715-9259-633215299810 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.182958] env[63531]: DEBUG oslo_vmware.api [None req-ead7921a-cb8c-490f-a3ab-414729b031bc tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Waiting for the task: (returnval){ [ 834.182958] env[63531]: value = "task-1117885" [ 834.182958] env[63531]: _type = "Task" [ 834.182958] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.191678] env[63531]: DEBUG oslo_vmware.api [None req-ead7921a-cb8c-490f-a3ab-414729b031bc tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Task: {'id': task-1117885, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.262235] env[63531]: DEBUG oslo_vmware.api [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Task: {'id': task-1117882, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.339872] env[63531]: DEBUG nova.network.neutron [req-5280cb6a-73a2-4b6f-86bf-aec5c72333e7 req-0ef86b50-799c-417b-8adb-20f3e060db04 service nova] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Updated VIF entry in instance network info cache for port b4141cf4-dab5-4033-8cc5-06966340323b. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 834.339872] env[63531]: DEBUG nova.network.neutron [req-5280cb6a-73a2-4b6f-86bf-aec5c72333e7 req-0ef86b50-799c-417b-8adb-20f3e060db04 service nova] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Updating instance_info_cache with network_info: [{"id": "b4141cf4-dab5-4033-8cc5-06966340323b", "address": "fa:16:3e:72:a7:59", "network": {"id": "7af8bbee-a43a-40b5-99c4-aee7e19b4093", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1560507257-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.195", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "68eaa455a44b4d96936fe8273629279a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ed82fc44-f216-40a7-a68b-62d76ffdb1f8", "external-id": "nsx-vlan-transportzone-236", "segmentation_id": 236, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4141cf4-da", "ovs_interfaceid": "b4141cf4-dab5-4033-8cc5-06966340323b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.393147] env[63531]: DEBUG oslo_vmware.api [None req-578f74de-4af2-4ff1-9a49-a39d5b6c11ae tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117883, 'name': PowerOnVM_Task, 'duration_secs': 0.441664} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.393515] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-578f74de-4af2-4ff1-9a49-a39d5b6c11ae tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 834.398021] env[63531]: DEBUG nova.compute.manager [None req-578f74de-4af2-4ff1-9a49-a39d5b6c11ae tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 834.399369] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90cb0bde-47a4-49b2-a99d-d0e524a27ebe {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.459747] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-92007411-3bac-410d-a9c6-c0b9a37679fa tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Creating linked-clone VM from snapshot {{(pid=63531) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 834.460516] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-24681956-f9b7-4bb6-9883-690cd9ae6b0a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.469374] env[63531]: DEBUG oslo_vmware.api [None req-92007411-3bac-410d-a9c6-c0b9a37679fa tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 834.469374] env[63531]: value = "task-1117886" [ 834.469374] env[63531]: _type = "Task" [ 834.469374] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.480445] env[63531]: DEBUG oslo_vmware.api [None req-92007411-3bac-410d-a9c6-c0b9a37679fa tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1117886, 'name': CloneVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.590397] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: c3920e60-b27f-496b-b616-afb23ab0bc26] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 834.597263] env[63531]: DEBUG nova.compute.utils [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 834.598795] env[63531]: DEBUG nova.compute.manager [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 834.598893] env[63531]: DEBUG nova.network.neutron [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 834.644297] env[63531]: DEBUG nova.policy [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '435c22cf5ad2496e90340da443ba4786', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '95a8149e13534b3faa6c70be2302bfce', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 834.694671] env[63531]: DEBUG oslo_vmware.api [None req-ead7921a-cb8c-490f-a3ab-414729b031bc tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Task: {'id': task-1117885, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.329485} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.695241] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-ead7921a-cb8c-490f-a3ab-414729b031bc tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 834.695623] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-ead7921a-cb8c-490f-a3ab-414729b031bc tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 834.695979] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-ead7921a-cb8c-490f-a3ab-414729b031bc tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 834.696353] env[63531]: INFO nova.compute.manager [None req-ead7921a-cb8c-490f-a3ab-414729b031bc tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Took 1.14 seconds to destroy the instance on the hypervisor. [ 834.696642] env[63531]: DEBUG oslo.service.loopingcall [None req-ead7921a-cb8c-490f-a3ab-414729b031bc tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 834.696859] env[63531]: DEBUG nova.compute.manager [-] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 834.696960] env[63531]: DEBUG nova.network.neutron [-] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 834.757407] env[63531]: DEBUG oslo_vmware.api [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Task: {'id': task-1117882, 'name': PowerOnVM_Task, 'duration_secs': 0.634896} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.757737] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] [instance: 43c29443-1210-4ee2-95c7-1257de308287] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 834.757943] env[63531]: INFO nova.compute.manager [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] [instance: 43c29443-1210-4ee2-95c7-1257de308287] Took 8.85 seconds to spawn the instance on the hypervisor. [ 834.758159] env[63531]: DEBUG nova.compute.manager [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] [instance: 43c29443-1210-4ee2-95c7-1257de308287] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 834.758967] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2eda3044-dfe9-40b9-8b4b-ed39524d6e7e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.844341] env[63531]: DEBUG oslo_concurrency.lockutils [req-5280cb6a-73a2-4b6f-86bf-aec5c72333e7 req-0ef86b50-799c-417b-8adb-20f3e060db04 service nova] Releasing lock "refresh_cache-0197e03d-6c36-4e73-9472-8e3319eb89ce" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.980155] env[63531]: DEBUG oslo_vmware.api [None req-92007411-3bac-410d-a9c6-c0b9a37679fa tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1117886, 'name': CloneVM_Task} progress is 94%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.025755] env[63531]: DEBUG nova.network.neutron [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] Successfully created port: e93a4c61-805f-4c14-8649-6718ca8f490d {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 835.093884] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: f39c8807-1e15-41cf-899b-a1fbe0695d58] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 835.102755] env[63531]: DEBUG nova.compute.manager [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 835.275253] env[63531]: INFO nova.compute.manager [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] [instance: 43c29443-1210-4ee2-95c7-1257de308287] Took 46.69 seconds to build instance. [ 835.398399] env[63531]: DEBUG nova.compute.manager [req-bec53a23-fe84-4f41-b2f7-6b690b7602f4 req-6d8eba93-b496-4dda-96de-391c42f06aba service nova] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Received event network-vif-deleted-b4141cf4-dab5-4033-8cc5-06966340323b {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 835.398769] env[63531]: INFO nova.compute.manager [req-bec53a23-fe84-4f41-b2f7-6b690b7602f4 req-6d8eba93-b496-4dda-96de-391c42f06aba service nova] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Neutron deleted interface b4141cf4-dab5-4033-8cc5-06966340323b; detaching it from the instance and deleting it from the info cache [ 835.399605] env[63531]: DEBUG nova.network.neutron [req-bec53a23-fe84-4f41-b2f7-6b690b7602f4 req-6d8eba93-b496-4dda-96de-391c42f06aba service nova] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.485621] env[63531]: DEBUG oslo_vmware.api [None req-92007411-3bac-410d-a9c6-c0b9a37679fa tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1117886, 'name': CloneVM_Task} progress is 94%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.549760] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d150eb79-9064-460f-a562-66a3f86ca26d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.560239] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0284bbf-5538-4675-b11f-fc2d00cad249 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.593348] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd50c040-5681-41c4-bb2a-de96085eaab9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.596839] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 645abb7a-cf73-4fb0-a9a0-49205f060d6f] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 835.602229] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df724b36-16e6-4eba-b4ab-4aee8fab97b8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.622889] env[63531]: DEBUG nova.compute.provider_tree [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 835.715209] env[63531]: DEBUG nova.network.neutron [-] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.778512] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a09c7145-7137-4190-b5e8-6b202247481b tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Lock "43c29443-1210-4ee2-95c7-1257de308287" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 74.516s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.902819] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b925e391-0caf-4300-8a56-3566a69f3ae9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.912662] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fe402ae-109a-4d2e-b58e-64e523ceb9e5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.945282] env[63531]: DEBUG nova.compute.manager [req-bec53a23-fe84-4f41-b2f7-6b690b7602f4 req-6d8eba93-b496-4dda-96de-391c42f06aba service nova] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Detach interface failed, port_id=b4141cf4-dab5-4033-8cc5-06966340323b, reason: Instance 0197e03d-6c36-4e73-9472-8e3319eb89ce could not be found. {{(pid=63531) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 835.981316] env[63531]: DEBUG oslo_vmware.api [None req-92007411-3bac-410d-a9c6-c0b9a37679fa tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1117886, 'name': CloneVM_Task} progress is 94%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.991180] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8f18d56a-00b5-41f4-ab3b-5e091e729fdb tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquiring lock "48043b5d-c0d3-4e8a-8d28-5a408d800e11" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.991929] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8f18d56a-00b5-41f4-ab3b-5e091e729fdb tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "48043b5d-c0d3-4e8a-8d28-5a408d800e11" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.108348] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: c3d597b4-1dc3-48d8-9bee-e73c8929181b] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 836.127675] env[63531]: DEBUG nova.compute.manager [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 836.130352] env[63531]: DEBUG nova.scheduler.client.report [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 836.156454] env[63531]: DEBUG nova.virt.hardware [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 836.156968] env[63531]: DEBUG nova.virt.hardware [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 836.157197] env[63531]: DEBUG nova.virt.hardware [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 836.157410] env[63531]: DEBUG nova.virt.hardware [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 836.157566] env[63531]: DEBUG nova.virt.hardware [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 836.157747] env[63531]: DEBUG nova.virt.hardware [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 836.158094] env[63531]: DEBUG nova.virt.hardware [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 836.158341] env[63531]: DEBUG nova.virt.hardware [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 836.158543] env[63531]: DEBUG nova.virt.hardware [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 836.158752] env[63531]: DEBUG nova.virt.hardware [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 836.158999] env[63531]: DEBUG nova.virt.hardware [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 836.160424] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-604ce1f0-a0f8-4238-b1e3-2363041595c7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.168955] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7d717a6-0970-480e-ade8-2b68ef7bffa7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.196109] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f49106e4-b7c8-427f-86d8-e698c2d39d08 tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Acquiring lock "43c29443-1210-4ee2-95c7-1257de308287" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.196377] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f49106e4-b7c8-427f-86d8-e698c2d39d08 tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Lock "43c29443-1210-4ee2-95c7-1257de308287" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.196627] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f49106e4-b7c8-427f-86d8-e698c2d39d08 tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Acquiring lock "43c29443-1210-4ee2-95c7-1257de308287-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.196805] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f49106e4-b7c8-427f-86d8-e698c2d39d08 tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Lock "43c29443-1210-4ee2-95c7-1257de308287-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.196994] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f49106e4-b7c8-427f-86d8-e698c2d39d08 tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Lock "43c29443-1210-4ee2-95c7-1257de308287-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.199736] env[63531]: INFO nova.compute.manager [None req-f49106e4-b7c8-427f-86d8-e698c2d39d08 tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] [instance: 43c29443-1210-4ee2-95c7-1257de308287] Terminating instance [ 836.202024] env[63531]: DEBUG nova.compute.manager [None req-f49106e4-b7c8-427f-86d8-e698c2d39d08 tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] [instance: 43c29443-1210-4ee2-95c7-1257de308287] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 836.202242] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f49106e4-b7c8-427f-86d8-e698c2d39d08 tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] [instance: 43c29443-1210-4ee2-95c7-1257de308287] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 836.203113] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37b6dc94-6fbc-492d-8b19-b3906d20f216 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.210468] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-f49106e4-b7c8-427f-86d8-e698c2d39d08 tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] [instance: 43c29443-1210-4ee2-95c7-1257de308287] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 836.210721] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bf8e3616-6d65-4c3b-bec7-820cb36c8e6e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.212418] env[63531]: INFO nova.compute.manager [None req-b1fa6345-cccc-4d75-a4d1-819aea9c59ef tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Unrescuing [ 836.212643] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b1fa6345-cccc-4d75-a4d1-819aea9c59ef tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Acquiring lock "refresh_cache-e0074e2e-e8bd-4bcb-9ad9-271a10e079e4" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.212793] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b1fa6345-cccc-4d75-a4d1-819aea9c59ef tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Acquired lock "refresh_cache-e0074e2e-e8bd-4bcb-9ad9-271a10e079e4" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.212957] env[63531]: DEBUG nova.network.neutron [None req-b1fa6345-cccc-4d75-a4d1-819aea9c59ef tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 836.217086] env[63531]: INFO nova.compute.manager [-] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Took 1.52 seconds to deallocate network for instance. [ 836.219864] env[63531]: DEBUG oslo_vmware.api [None req-f49106e4-b7c8-427f-86d8-e698c2d39d08 tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Waiting for the task: (returnval){ [ 836.219864] env[63531]: value = "task-1117887" [ 836.219864] env[63531]: _type = "Task" [ 836.219864] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.234358] env[63531]: DEBUG oslo_vmware.api [None req-f49106e4-b7c8-427f-86d8-e698c2d39d08 tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Task: {'id': task-1117887, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.281240] env[63531]: DEBUG nova.compute.manager [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 836.485168] env[63531]: DEBUG oslo_vmware.api [None req-92007411-3bac-410d-a9c6-c0b9a37679fa tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1117886, 'name': CloneVM_Task, 'duration_secs': 1.920219} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.485613] env[63531]: INFO nova.virt.vmwareapi.vmops [None req-92007411-3bac-410d-a9c6-c0b9a37679fa tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Created linked-clone VM from snapshot [ 836.486624] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e076e8af-7356-435d-a24f-8280cc6a5a0c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.498246] env[63531]: DEBUG nova.compute.utils [None req-8f18d56a-00b5-41f4-ab3b-5e091e729fdb tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 836.499554] env[63531]: DEBUG nova.virt.vmwareapi.images [None req-92007411-3bac-410d-a9c6-c0b9a37679fa tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Uploading image 4f765107-1431-4888-adf7-e8b9dbcda251 {{(pid=63531) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 836.521647] env[63531]: DEBUG oslo_vmware.rw_handles [None req-92007411-3bac-410d-a9c6-c0b9a37679fa tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 836.521647] env[63531]: value = "vm-244694" [ 836.521647] env[63531]: _type = "VirtualMachine" [ 836.521647] env[63531]: }. {{(pid=63531) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 836.521995] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-2600b80c-7685-448a-9920-fffb16e0e9a2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.529624] env[63531]: DEBUG oslo_vmware.rw_handles [None req-92007411-3bac-410d-a9c6-c0b9a37679fa tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lease: (returnval){ [ 836.529624] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52455b4f-1b4d-75c8-b6eb-e866a9d05c65" [ 836.529624] env[63531]: _type = "HttpNfcLease" [ 836.529624] env[63531]: } obtained for exporting VM: (result){ [ 836.529624] env[63531]: value = "vm-244694" [ 836.529624] env[63531]: _type = "VirtualMachine" [ 836.529624] env[63531]: }. {{(pid=63531) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 836.529624] env[63531]: DEBUG oslo_vmware.api [None req-92007411-3bac-410d-a9c6-c0b9a37679fa tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the lease: (returnval){ [ 836.529624] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52455b4f-1b4d-75c8-b6eb-e866a9d05c65" [ 836.529624] env[63531]: _type = "HttpNfcLease" [ 836.529624] env[63531]: } to be ready. {{(pid=63531) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 836.537642] env[63531]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 836.537642] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52455b4f-1b4d-75c8-b6eb-e866a9d05c65" [ 836.537642] env[63531]: _type = "HttpNfcLease" [ 836.537642] env[63531]: } is initializing. {{(pid=63531) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 836.611106] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: f0276e34-1de3-482a-82c7-1439d40fd85e] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 836.635737] env[63531]: DEBUG oslo_concurrency.lockutils [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.546s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.636351] env[63531]: DEBUG nova.compute.manager [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] [instance: 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 836.642225] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 35.665s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.642225] env[63531]: INFO nova.compute.claims [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 3eeaa6f9-60d9-43fe-a438-fbf2ceececae] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 836.728061] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ead7921a-cb8c-490f-a3ab-414729b031bc tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.733872] env[63531]: DEBUG oslo_vmware.api [None req-f49106e4-b7c8-427f-86d8-e698c2d39d08 tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Task: {'id': task-1117887, 'name': PowerOffVM_Task, 'duration_secs': 0.221762} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.734137] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-f49106e4-b7c8-427f-86d8-e698c2d39d08 tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] [instance: 43c29443-1210-4ee2-95c7-1257de308287] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 836.734333] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f49106e4-b7c8-427f-86d8-e698c2d39d08 tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] [instance: 43c29443-1210-4ee2-95c7-1257de308287] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 836.734595] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bbec2fc1-070e-469d-a29b-0e6c67f3cfaf {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.803786] env[63531]: DEBUG oslo_concurrency.lockutils [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.838718] env[63531]: DEBUG nova.network.neutron [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] Successfully updated port: e93a4c61-805f-4c14-8649-6718ca8f490d {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 836.897692] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f49106e4-b7c8-427f-86d8-e698c2d39d08 tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] [instance: 43c29443-1210-4ee2-95c7-1257de308287] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 836.898094] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f49106e4-b7c8-427f-86d8-e698c2d39d08 tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] [instance: 43c29443-1210-4ee2-95c7-1257de308287] Deleting contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 836.898411] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-f49106e4-b7c8-427f-86d8-e698c2d39d08 tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Deleting the datastore file [datastore2] 43c29443-1210-4ee2-95c7-1257de308287 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 836.898825] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e69fd373-b9a4-479c-982a-8ee421586e60 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.906763] env[63531]: DEBUG oslo_vmware.api [None req-f49106e4-b7c8-427f-86d8-e698c2d39d08 tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Waiting for the task: (returnval){ [ 836.906763] env[63531]: value = "task-1117890" [ 836.906763] env[63531]: _type = "Task" [ 836.906763] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.916185] env[63531]: DEBUG oslo_vmware.api [None req-f49106e4-b7c8-427f-86d8-e698c2d39d08 tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Task: {'id': task-1117890, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.933328] env[63531]: DEBUG nova.network.neutron [None req-b1fa6345-cccc-4d75-a4d1-819aea9c59ef tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Updating instance_info_cache with network_info: [{"id": "f1ef6f0e-1f62-4034-8edb-e8cc1bb98abd", "address": "fa:16:3e:3e:82:e5", "network": {"id": "dc416058-1c04-47ac-8b73-1557f7c33d1c", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1350689703-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "824ed405cc6e4868a70e82489e05e8ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "05b1253d-2b87-4158-9ff1-dafcf829f11f", "external-id": "nsx-vlan-transportzone-55", "segmentation_id": 55, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf1ef6f0e-1f", "ovs_interfaceid": "f1ef6f0e-1f62-4034-8edb-e8cc1bb98abd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 837.003407] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8f18d56a-00b5-41f4-ab3b-5e091e729fdb tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "48043b5d-c0d3-4e8a-8d28-5a408d800e11" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.012s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.010454] env[63531]: DEBUG nova.compute.manager [req-cecdf7cc-4c34-4bd5-b706-9caaa24f087d req-131c4436-ea46-47e1-b391-f830651dfc3e service nova] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] Received event network-vif-plugged-e93a4c61-805f-4c14-8649-6718ca8f490d {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 837.010746] env[63531]: DEBUG oslo_concurrency.lockutils [req-cecdf7cc-4c34-4bd5-b706-9caaa24f087d req-131c4436-ea46-47e1-b391-f830651dfc3e service nova] Acquiring lock "45b05c54-6865-47a1-ba93-90ad3e1ba07e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 837.010993] env[63531]: DEBUG oslo_concurrency.lockutils [req-cecdf7cc-4c34-4bd5-b706-9caaa24f087d req-131c4436-ea46-47e1-b391-f830651dfc3e service nova] Lock "45b05c54-6865-47a1-ba93-90ad3e1ba07e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.011284] env[63531]: DEBUG oslo_concurrency.lockutils [req-cecdf7cc-4c34-4bd5-b706-9caaa24f087d req-131c4436-ea46-47e1-b391-f830651dfc3e service nova] Lock "45b05c54-6865-47a1-ba93-90ad3e1ba07e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.011500] env[63531]: DEBUG nova.compute.manager [req-cecdf7cc-4c34-4bd5-b706-9caaa24f087d req-131c4436-ea46-47e1-b391-f830651dfc3e service nova] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] No waiting events found dispatching network-vif-plugged-e93a4c61-805f-4c14-8649-6718ca8f490d {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 837.011714] env[63531]: WARNING nova.compute.manager [req-cecdf7cc-4c34-4bd5-b706-9caaa24f087d req-131c4436-ea46-47e1-b391-f830651dfc3e service nova] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] Received unexpected event network-vif-plugged-e93a4c61-805f-4c14-8649-6718ca8f490d for instance with vm_state building and task_state spawning. [ 837.038159] env[63531]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 837.038159] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52455b4f-1b4d-75c8-b6eb-e866a9d05c65" [ 837.038159] env[63531]: _type = "HttpNfcLease" [ 837.038159] env[63531]: } is ready. {{(pid=63531) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 837.039045] env[63531]: DEBUG oslo_vmware.rw_handles [None req-92007411-3bac-410d-a9c6-c0b9a37679fa tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 837.039045] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52455b4f-1b4d-75c8-b6eb-e866a9d05c65" [ 837.039045] env[63531]: _type = "HttpNfcLease" [ 837.039045] env[63531]: }. {{(pid=63531) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 837.039420] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f233129e-e6f9-4faf-92bc-d6e3a3d6458b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.048253] env[63531]: DEBUG oslo_vmware.rw_handles [None req-92007411-3bac-410d-a9c6-c0b9a37679fa tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f3461f-e2c4-1251-e983-2cc3133ddf8a/disk-0.vmdk from lease info. {{(pid=63531) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 837.048534] env[63531]: DEBUG oslo_vmware.rw_handles [None req-92007411-3bac-410d-a9c6-c0b9a37679fa tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f3461f-e2c4-1251-e983-2cc3133ddf8a/disk-0.vmdk for reading. {{(pid=63531) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 837.114101] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 7bc63fdd-4b50-486d-9681-ca0baa08f7b3] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 837.150063] env[63531]: DEBUG nova.compute.utils [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 837.154129] env[63531]: DEBUG nova.compute.manager [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] [instance: 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 837.155117] env[63531]: DEBUG nova.network.neutron [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] [instance: 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 837.158463] env[63531]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-06ac7e54-7626-464e-a3d8-c5173ab315b3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.199775] env[63531]: DEBUG nova.policy [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8d2abbf7198f410eacf9c8297015af17', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3e47ec69a0304ada957bbbcbf881460e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 837.341355] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Acquiring lock "refresh_cache-45b05c54-6865-47a1-ba93-90ad3e1ba07e" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.341644] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Acquired lock "refresh_cache-45b05c54-6865-47a1-ba93-90ad3e1ba07e" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.341746] env[63531]: DEBUG nova.network.neutron [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 837.420215] env[63531]: DEBUG oslo_vmware.api [None req-f49106e4-b7c8-427f-86d8-e698c2d39d08 tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Task: {'id': task-1117890, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.293366} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.421100] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-f49106e4-b7c8-427f-86d8-e698c2d39d08 tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 837.421374] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f49106e4-b7c8-427f-86d8-e698c2d39d08 tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] [instance: 43c29443-1210-4ee2-95c7-1257de308287] Deleted contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 837.421644] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f49106e4-b7c8-427f-86d8-e698c2d39d08 tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] [instance: 43c29443-1210-4ee2-95c7-1257de308287] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 837.422036] env[63531]: INFO nova.compute.manager [None req-f49106e4-b7c8-427f-86d8-e698c2d39d08 tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] [instance: 43c29443-1210-4ee2-95c7-1257de308287] Took 1.22 seconds to destroy the instance on the hypervisor. [ 837.422374] env[63531]: DEBUG oslo.service.loopingcall [None req-f49106e4-b7c8-427f-86d8-e698c2d39d08 tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 837.423229] env[63531]: DEBUG nova.compute.manager [-] [instance: 43c29443-1210-4ee2-95c7-1257de308287] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 837.423400] env[63531]: DEBUG nova.network.neutron [-] [instance: 43c29443-1210-4ee2-95c7-1257de308287] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 837.438530] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b1fa6345-cccc-4d75-a4d1-819aea9c59ef tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Releasing lock "refresh_cache-e0074e2e-e8bd-4bcb-9ad9-271a10e079e4" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.438530] env[63531]: DEBUG nova.objects.instance [None req-b1fa6345-cccc-4d75-a4d1-819aea9c59ef tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Lazy-loading 'flavor' on Instance uuid e0074e2e-e8bd-4bcb-9ad9-271a10e079e4 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 837.466917] env[63531]: DEBUG nova.network.neutron [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] [instance: 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d] Successfully created port: f7815862-8080-412a-8d54-7bad36f1413b {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 837.618549] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 0c948655-362d-4673-b3e6-3f28ec69ea3c] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 837.658313] env[63531]: DEBUG nova.compute.manager [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] [instance: 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 837.892753] env[63531]: DEBUG nova.network.neutron [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 837.946610] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d85fdd88-31aa-4a52-8b10-84ae2ef65047 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.972918] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1fa6345-cccc-4d75-a4d1-819aea9c59ef tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 837.979291] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8fbae1fd-a9c3-4f98-9207-7c06e3b41589 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.986897] env[63531]: DEBUG oslo_vmware.api [None req-b1fa6345-cccc-4d75-a4d1-819aea9c59ef tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Waiting for the task: (returnval){ [ 837.986897] env[63531]: value = "task-1117891" [ 837.986897] env[63531]: _type = "Task" [ 837.986897] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.995885] env[63531]: DEBUG oslo_vmware.api [None req-b1fa6345-cccc-4d75-a4d1-819aea9c59ef tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117891, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.077148] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8f18d56a-00b5-41f4-ab3b-5e091e729fdb tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquiring lock "48043b5d-c0d3-4e8a-8d28-5a408d800e11" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 838.078952] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8f18d56a-00b5-41f4-ab3b-5e091e729fdb tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "48043b5d-c0d3-4e8a-8d28-5a408d800e11" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 838.078952] env[63531]: INFO nova.compute.manager [None req-8f18d56a-00b5-41f4-ab3b-5e091e729fdb tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Attaching volume 02e72028-7ac8-4eab-88b0-9de83339ce64 to /dev/sdb [ 838.125351] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: ffb34f90-d22f-440b-ba07-75d474c3c300] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 838.125351] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-739fed1e-b7c9-46bc-b1e9-e3771168f881 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.136357] env[63531]: DEBUG nova.network.neutron [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] Updating instance_info_cache with network_info: [{"id": "e93a4c61-805f-4c14-8649-6718ca8f490d", "address": "fa:16:3e:63:6d:f8", "network": {"id": "4dfee301-fd80-46f1-8ed0-0499137e0bf1", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-780216327-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "95a8149e13534b3faa6c70be2302bfce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "29945b49-d2b8-449d-9531-437917f49839", "external-id": "nsx-vlan-transportzone-787", "segmentation_id": 787, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape93a4c61-80", "ovs_interfaceid": "e93a4c61-805f-4c14-8649-6718ca8f490d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 838.144766] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-984e2f04-7680-4e53-ad43-a2ad159b1921 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.159515] env[63531]: DEBUG nova.virt.block_device [None req-8f18d56a-00b5-41f4-ab3b-5e091e729fdb tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Updating existing volume attachment record: b09719bf-d11f-4955-a9cc-3c0f536f446a {{(pid=63531) _volume_attach /opt/stack/nova/nova/virt/block_device.py:679}} [ 838.167731] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dff4bdc-8e84-4028-89f3-b5de68d901aa {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.177508] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f22a9854-f855-4075-a709-fb2421c9d738 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.217097] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3165c287-a3af-4392-bcb9-5d955de703ab {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.226183] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4add80ae-b54a-41e9-8138-87118b0d7104 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.241644] env[63531]: DEBUG nova.compute.provider_tree [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 838.281356] env[63531]: DEBUG nova.network.neutron [-] [instance: 43c29443-1210-4ee2-95c7-1257de308287] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 838.498531] env[63531]: DEBUG oslo_vmware.api [None req-b1fa6345-cccc-4d75-a4d1-819aea9c59ef tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117891, 'name': PowerOffVM_Task, 'duration_secs': 0.212953} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.498997] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1fa6345-cccc-4d75-a4d1-819aea9c59ef tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 838.505541] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-b1fa6345-cccc-4d75-a4d1-819aea9c59ef tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Reconfiguring VM instance instance-0000003d to detach disk 2001 {{(pid=63531) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 838.506008] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8e59fa60-ad59-4512-be60-9d1eca7fc1f7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.525610] env[63531]: DEBUG oslo_vmware.api [None req-b1fa6345-cccc-4d75-a4d1-819aea9c59ef tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Waiting for the task: (returnval){ [ 838.525610] env[63531]: value = "task-1117895" [ 838.525610] env[63531]: _type = "Task" [ 838.525610] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.536957] env[63531]: DEBUG oslo_vmware.api [None req-b1fa6345-cccc-4d75-a4d1-819aea9c59ef tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117895, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.633407] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 5b86b63a-f6c1-4983-a3cc-3cde7a10d17b] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 838.640050] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Releasing lock "refresh_cache-45b05c54-6865-47a1-ba93-90ad3e1ba07e" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.640379] env[63531]: DEBUG nova.compute.manager [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] Instance network_info: |[{"id": "e93a4c61-805f-4c14-8649-6718ca8f490d", "address": "fa:16:3e:63:6d:f8", "network": {"id": "4dfee301-fd80-46f1-8ed0-0499137e0bf1", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-780216327-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "95a8149e13534b3faa6c70be2302bfce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "29945b49-d2b8-449d-9531-437917f49839", "external-id": "nsx-vlan-transportzone-787", "segmentation_id": 787, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape93a4c61-80", "ovs_interfaceid": "e93a4c61-805f-4c14-8649-6718ca8f490d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 838.640845] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:63:6d:f8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '29945b49-d2b8-449d-9531-437917f49839', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e93a4c61-805f-4c14-8649-6718ca8f490d', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 838.649018] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Creating folder: Project (95a8149e13534b3faa6c70be2302bfce). Parent ref: group-v244585. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 838.649578] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-19579044-5f4e-41e0-a28b-b24bc231f9c0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.661676] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Created folder: Project (95a8149e13534b3faa6c70be2302bfce) in parent group-v244585. [ 838.661937] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Creating folder: Instances. Parent ref: group-v244697. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 838.662312] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-925617c4-cec9-4943-957b-9931302101d4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.673181] env[63531]: DEBUG nova.compute.manager [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] [instance: 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 838.675184] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Created folder: Instances in parent group-v244697. [ 838.675477] env[63531]: DEBUG oslo.service.loopingcall [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 838.675987] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 838.676327] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-04a80c09-8cf8-40f3-abec-0c60da334d0a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.698466] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 838.698466] env[63531]: value = "task-1117898" [ 838.698466] env[63531]: _type = "Task" [ 838.698466] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.707680] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117898, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.745453] env[63531]: DEBUG nova.scheduler.client.report [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 838.785322] env[63531]: INFO nova.compute.manager [-] [instance: 43c29443-1210-4ee2-95c7-1257de308287] Took 1.36 seconds to deallocate network for instance. [ 839.037234] env[63531]: DEBUG oslo_vmware.api [None req-b1fa6345-cccc-4d75-a4d1-819aea9c59ef tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117895, 'name': ReconfigVM_Task, 'duration_secs': 0.242348} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.038188] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-b1fa6345-cccc-4d75-a4d1-819aea9c59ef tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Reconfigured VM instance instance-0000003d to detach disk 2001 {{(pid=63531) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 839.038188] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1fa6345-cccc-4d75-a4d1-819aea9c59ef tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 839.038921] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dd527126-d38c-4af3-a9d2-0cee7048df3d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.046082] env[63531]: DEBUG oslo_vmware.api [None req-b1fa6345-cccc-4d75-a4d1-819aea9c59ef tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Waiting for the task: (returnval){ [ 839.046082] env[63531]: value = "task-1117899" [ 839.046082] env[63531]: _type = "Task" [ 839.046082] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.054553] env[63531]: DEBUG oslo_vmware.api [None req-b1fa6345-cccc-4d75-a4d1-819aea9c59ef tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117899, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.114728] env[63531]: DEBUG nova.compute.manager [req-2fc299ee-ed5a-4fe0-9a73-4302ec47b8fe req-b42e9e2f-9af0-4a4e-83ed-1215b987d5b3 service nova] [instance: 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d] Received event network-vif-plugged-f7815862-8080-412a-8d54-7bad36f1413b {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 839.114957] env[63531]: DEBUG oslo_concurrency.lockutils [req-2fc299ee-ed5a-4fe0-9a73-4302ec47b8fe req-b42e9e2f-9af0-4a4e-83ed-1215b987d5b3 service nova] Acquiring lock "56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.115215] env[63531]: DEBUG oslo_concurrency.lockutils [req-2fc299ee-ed5a-4fe0-9a73-4302ec47b8fe req-b42e9e2f-9af0-4a4e-83ed-1215b987d5b3 service nova] Lock "56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.115487] env[63531]: DEBUG oslo_concurrency.lockutils [req-2fc299ee-ed5a-4fe0-9a73-4302ec47b8fe req-b42e9e2f-9af0-4a4e-83ed-1215b987d5b3 service nova] Lock "56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.115591] env[63531]: DEBUG nova.compute.manager [req-2fc299ee-ed5a-4fe0-9a73-4302ec47b8fe req-b42e9e2f-9af0-4a4e-83ed-1215b987d5b3 service nova] [instance: 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d] No waiting events found dispatching network-vif-plugged-f7815862-8080-412a-8d54-7bad36f1413b {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 839.115793] env[63531]: WARNING nova.compute.manager [req-2fc299ee-ed5a-4fe0-9a73-4302ec47b8fe req-b42e9e2f-9af0-4a4e-83ed-1215b987d5b3 service nova] [instance: 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d] Received unexpected event network-vif-plugged-f7815862-8080-412a-8d54-7bad36f1413b for instance with vm_state building and task_state spawning. [ 839.131273] env[63531]: DEBUG nova.network.neutron [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] [instance: 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d] Successfully updated port: f7815862-8080-412a-8d54-7bad36f1413b {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 839.137649] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 4410b013-868a-41e7-9dfb-584fb2d9d142] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 839.192269] env[63531]: DEBUG nova.compute.manager [req-7704704a-3a1a-4dc8-9b72-48be6b9f77e6 req-d85b6242-1fb3-4b2a-8e6f-6895cf87738b service nova] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] Received event network-changed-e93a4c61-805f-4c14-8649-6718ca8f490d {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 839.192462] env[63531]: DEBUG nova.compute.manager [req-7704704a-3a1a-4dc8-9b72-48be6b9f77e6 req-d85b6242-1fb3-4b2a-8e6f-6895cf87738b service nova] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] Refreshing instance network info cache due to event network-changed-e93a4c61-805f-4c14-8649-6718ca8f490d. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 839.192688] env[63531]: DEBUG oslo_concurrency.lockutils [req-7704704a-3a1a-4dc8-9b72-48be6b9f77e6 req-d85b6242-1fb3-4b2a-8e6f-6895cf87738b service nova] Acquiring lock "refresh_cache-45b05c54-6865-47a1-ba93-90ad3e1ba07e" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.192835] env[63531]: DEBUG oslo_concurrency.lockutils [req-7704704a-3a1a-4dc8-9b72-48be6b9f77e6 req-d85b6242-1fb3-4b2a-8e6f-6895cf87738b service nova] Acquired lock "refresh_cache-45b05c54-6865-47a1-ba93-90ad3e1ba07e" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.192999] env[63531]: DEBUG nova.network.neutron [req-7704704a-3a1a-4dc8-9b72-48be6b9f77e6 req-d85b6242-1fb3-4b2a-8e6f-6895cf87738b service nova] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] Refreshing network info cache for port e93a4c61-805f-4c14-8649-6718ca8f490d {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 839.209749] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117898, 'name': CreateVM_Task, 'duration_secs': 0.371086} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.209972] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 839.210723] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.211519] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.211911] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 839.212509] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f6838fbc-bb19-4be2-9ba4-41a0ddb4094f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.218076] env[63531]: DEBUG oslo_vmware.api [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Waiting for the task: (returnval){ [ 839.218076] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52ff6bb8-c47a-b588-9757-1ca9098980fd" [ 839.218076] env[63531]: _type = "Task" [ 839.218076] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.226287] env[63531]: DEBUG oslo_vmware.api [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52ff6bb8-c47a-b588-9757-1ca9098980fd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.251653] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.612s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.252216] env[63531]: DEBUG nova.compute.manager [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 3eeaa6f9-60d9-43fe-a438-fbf2ceececae] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 839.255694] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.461s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.257067] env[63531]: INFO nova.compute.claims [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 839.292267] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f49106e4-b7c8-427f-86d8-e698c2d39d08 tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.521083] env[63531]: DEBUG nova.virt.hardware [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 839.521459] env[63531]: DEBUG nova.virt.hardware [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 839.521684] env[63531]: DEBUG nova.virt.hardware [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 839.521905] env[63531]: DEBUG nova.virt.hardware [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 839.522167] env[63531]: DEBUG nova.virt.hardware [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 839.522377] env[63531]: DEBUG nova.virt.hardware [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 839.522657] env[63531]: DEBUG nova.virt.hardware [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 839.522906] env[63531]: DEBUG nova.virt.hardware [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 839.523191] env[63531]: DEBUG nova.virt.hardware [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 839.523416] env[63531]: DEBUG nova.virt.hardware [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 839.523646] env[63531]: DEBUG nova.virt.hardware [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 839.526239] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d454977-1959-4788-b7ad-c420b4e7036f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.536422] env[63531]: DEBUG oslo_vmware.rw_handles [None req-13b8019c-692f-4097-a7ee-b3251d4a991c tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524335f2-851d-a3af-b6a7-97a78a8f2d6e/disk-0.vmdk. {{(pid=63531) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 839.538009] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ca0bf69-a1ca-4b4f-8e20-cb1de9eecba1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.544920] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c45f65a8-c38b-459f-a592-90cc700ff1b8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.563886] env[63531]: DEBUG oslo_vmware.rw_handles [None req-13b8019c-692f-4097-a7ee-b3251d4a991c tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524335f2-851d-a3af-b6a7-97a78a8f2d6e/disk-0.vmdk is in state: ready. {{(pid=63531) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 839.564099] env[63531]: ERROR oslo_vmware.rw_handles [None req-13b8019c-692f-4097-a7ee-b3251d4a991c tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524335f2-851d-a3af-b6a7-97a78a8f2d6e/disk-0.vmdk due to incomplete transfer. [ 839.564376] env[63531]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-5184f2db-6e2e-4795-aa5d-7c2118502df1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.568981] env[63531]: DEBUG oslo_vmware.api [None req-b1fa6345-cccc-4d75-a4d1-819aea9c59ef tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117899, 'name': PowerOnVM_Task, 'duration_secs': 0.417106} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.569601] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1fa6345-cccc-4d75-a4d1-819aea9c59ef tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 839.569844] env[63531]: DEBUG nova.compute.manager [None req-b1fa6345-cccc-4d75-a4d1-819aea9c59ef tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 839.570678] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37a93749-9197-41ce-89e2-ce6d90bc2452 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.574736] env[63531]: DEBUG oslo_vmware.rw_handles [None req-13b8019c-692f-4097-a7ee-b3251d4a991c tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524335f2-851d-a3af-b6a7-97a78a8f2d6e/disk-0.vmdk. {{(pid=63531) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 839.574988] env[63531]: DEBUG nova.virt.vmwareapi.images [None req-13b8019c-692f-4097-a7ee-b3251d4a991c tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Uploaded image 2ef8015f-f209-49ec-9140-1466c4081402 to the Glance image server {{(pid=63531) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 839.577491] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-13b8019c-692f-4097-a7ee-b3251d4a991c tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Destroying the VM {{(pid=63531) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 839.579202] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-27c34848-0be4-45c1-8a50-3befacafe773 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.587730] env[63531]: DEBUG oslo_vmware.api [None req-13b8019c-692f-4097-a7ee-b3251d4a991c tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Waiting for the task: (returnval){ [ 839.587730] env[63531]: value = "task-1117900" [ 839.587730] env[63531]: _type = "Task" [ 839.587730] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.596314] env[63531]: DEBUG oslo_vmware.api [None req-13b8019c-692f-4097-a7ee-b3251d4a991c tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117900, 'name': Destroy_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.634250] env[63531]: DEBUG oslo_concurrency.lockutils [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Acquiring lock "refresh_cache-56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.634683] env[63531]: DEBUG oslo_concurrency.lockutils [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Acquired lock "refresh_cache-56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.634683] env[63531]: DEBUG nova.network.neutron [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] [instance: 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 839.640546] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 839.640760] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Cleaning up deleted instances with incomplete migration {{(pid=63531) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 839.728192] env[63531]: DEBUG oslo_vmware.api [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52ff6bb8-c47a-b588-9757-1ca9098980fd, 'name': SearchDatastore_Task, 'duration_secs': 0.011119} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.732160] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.732160] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 839.732160] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.732160] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.732160] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 839.732160] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-905db290-c9e7-49a2-9754-4fcfb8611f2f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.741728] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 839.741910] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 839.742625] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-044e54f9-4569-40c4-9f1a-3ce7b1a1c6c5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.747761] env[63531]: DEBUG oslo_vmware.api [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Waiting for the task: (returnval){ [ 839.747761] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]522bfe6e-2ec0-248a-c78e-25fbd0965526" [ 839.747761] env[63531]: _type = "Task" [ 839.747761] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.757777] env[63531]: DEBUG nova.compute.utils [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 839.759582] env[63531]: DEBUG oslo_vmware.api [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]522bfe6e-2ec0-248a-c78e-25fbd0965526, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.759888] env[63531]: DEBUG nova.compute.manager [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 3eeaa6f9-60d9-43fe-a438-fbf2ceececae] Not allocating networking since 'none' was specified. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 839.929570] env[63531]: DEBUG nova.network.neutron [req-7704704a-3a1a-4dc8-9b72-48be6b9f77e6 req-d85b6242-1fb3-4b2a-8e6f-6895cf87738b service nova] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] Updated VIF entry in instance network info cache for port e93a4c61-805f-4c14-8649-6718ca8f490d. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 839.929979] env[63531]: DEBUG nova.network.neutron [req-7704704a-3a1a-4dc8-9b72-48be6b9f77e6 req-d85b6242-1fb3-4b2a-8e6f-6895cf87738b service nova] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] Updating instance_info_cache with network_info: [{"id": "e93a4c61-805f-4c14-8649-6718ca8f490d", "address": "fa:16:3e:63:6d:f8", "network": {"id": "4dfee301-fd80-46f1-8ed0-0499137e0bf1", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-780216327-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "95a8149e13534b3faa6c70be2302bfce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "29945b49-d2b8-449d-9531-437917f49839", "external-id": "nsx-vlan-transportzone-787", "segmentation_id": 787, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape93a4c61-80", "ovs_interfaceid": "e93a4c61-805f-4c14-8649-6718ca8f490d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 840.097380] env[63531]: DEBUG oslo_vmware.api [None req-13b8019c-692f-4097-a7ee-b3251d4a991c tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117900, 'name': Destroy_Task, 'duration_secs': 0.45787} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.097668] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-13b8019c-692f-4097-a7ee-b3251d4a991c tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Destroyed the VM [ 840.097917] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-13b8019c-692f-4097-a7ee-b3251d4a991c tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Deleting Snapshot of the VM instance {{(pid=63531) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 840.098171] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-7a434eec-c169-428d-904a-31636c4a5b1a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.105158] env[63531]: DEBUG oslo_vmware.api [None req-13b8019c-692f-4097-a7ee-b3251d4a991c tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Waiting for the task: (returnval){ [ 840.105158] env[63531]: value = "task-1117901" [ 840.105158] env[63531]: _type = "Task" [ 840.105158] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.113133] env[63531]: DEBUG oslo_vmware.api [None req-13b8019c-692f-4097-a7ee-b3251d4a991c tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117901, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.143124] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 840.169445] env[63531]: DEBUG nova.network.neutron [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] [instance: 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 840.257555] env[63531]: DEBUG oslo_vmware.api [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]522bfe6e-2ec0-248a-c78e-25fbd0965526, 'name': SearchDatastore_Task, 'duration_secs': 0.015524} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.258379] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1678a412-bafd-4766-bb9e-e6d3c6732050 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.264924] env[63531]: DEBUG nova.compute.manager [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 3eeaa6f9-60d9-43fe-a438-fbf2ceececae] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 840.268099] env[63531]: DEBUG oslo_vmware.api [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Waiting for the task: (returnval){ [ 840.268099] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]520cb5b7-d9a3-ee67-b874-e9f078caa632" [ 840.268099] env[63531]: _type = "Task" [ 840.268099] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.283285] env[63531]: DEBUG oslo_vmware.api [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]520cb5b7-d9a3-ee67-b874-e9f078caa632, 'name': SearchDatastore_Task, 'duration_secs': 0.015682} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.283561] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 840.283999] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 45b05c54-6865-47a1-ba93-90ad3e1ba07e/45b05c54-6865-47a1-ba93-90ad3e1ba07e.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 840.284123] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-17e4fada-3530-41b8-8009-5684799a0dfd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.291157] env[63531]: DEBUG oslo_vmware.api [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Waiting for the task: (returnval){ [ 840.291157] env[63531]: value = "task-1117902" [ 840.291157] env[63531]: _type = "Task" [ 840.291157] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.304017] env[63531]: DEBUG oslo_vmware.api [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Task: {'id': task-1117902, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.315675] env[63531]: DEBUG nova.network.neutron [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] [instance: 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d] Updating instance_info_cache with network_info: [{"id": "f7815862-8080-412a-8d54-7bad36f1413b", "address": "fa:16:3e:fd:cf:42", "network": {"id": "b47d3a83-c039-4eb2-b9b1-50898df74a35", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-465774433-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3e47ec69a0304ada957bbbcbf881460e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "da0e5087-d65b-416f-90fe-beaa9c534ad3", "external-id": "nsx-vlan-transportzone-522", "segmentation_id": 522, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf7815862-80", "ovs_interfaceid": "f7815862-8080-412a-8d54-7bad36f1413b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 840.433745] env[63531]: DEBUG oslo_concurrency.lockutils [req-7704704a-3a1a-4dc8-9b72-48be6b9f77e6 req-d85b6242-1fb3-4b2a-8e6f-6895cf87738b service nova] Releasing lock "refresh_cache-45b05c54-6865-47a1-ba93-90ad3e1ba07e" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 840.434214] env[63531]: DEBUG nova.compute.manager [req-7704704a-3a1a-4dc8-9b72-48be6b9f77e6 req-d85b6242-1fb3-4b2a-8e6f-6895cf87738b service nova] [instance: 43c29443-1210-4ee2-95c7-1257de308287] Received event network-vif-deleted-73fd445a-3029-48b0-b3d7-7246dc21081b {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 840.613909] env[63531]: DEBUG oslo_vmware.api [None req-13b8019c-692f-4097-a7ee-b3251d4a991c tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117901, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.616017] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa3b1839-217c-4e71-a80d-2db5d550cabf {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.622328] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aceab581-7e6e-4504-bc37-ac28dfbde536 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.652040] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d12608d-cb92-47a7-9339-3ef07337fe9b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.661046] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b828f973-976c-47e5-a9eb-286c2ffbdbcd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.674353] env[63531]: DEBUG nova.compute.provider_tree [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 840.801499] env[63531]: DEBUG oslo_vmware.api [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Task: {'id': task-1117902, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.819031] env[63531]: DEBUG oslo_concurrency.lockutils [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Releasing lock "refresh_cache-56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 840.819031] env[63531]: DEBUG nova.compute.manager [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] [instance: 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d] Instance network_info: |[{"id": "f7815862-8080-412a-8d54-7bad36f1413b", "address": "fa:16:3e:fd:cf:42", "network": {"id": "b47d3a83-c039-4eb2-b9b1-50898df74a35", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-465774433-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3e47ec69a0304ada957bbbcbf881460e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "da0e5087-d65b-416f-90fe-beaa9c534ad3", "external-id": "nsx-vlan-transportzone-522", "segmentation_id": 522, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf7815862-80", "ovs_interfaceid": "f7815862-8080-412a-8d54-7bad36f1413b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 840.819821] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] [instance: 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fd:cf:42', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'da0e5087-d65b-416f-90fe-beaa9c534ad3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f7815862-8080-412a-8d54-7bad36f1413b', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 840.829556] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Creating folder: Project (3e47ec69a0304ada957bbbcbf881460e). Parent ref: group-v244585. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 840.829901] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-422c4cb2-85c5-42c3-83ca-fce427e9d9e4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.839803] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Created folder: Project (3e47ec69a0304ada957bbbcbf881460e) in parent group-v244585. [ 840.840157] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Creating folder: Instances. Parent ref: group-v244700. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 840.840522] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-71c70130-4054-4473-acac-a391b2eb47ed {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.849811] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Created folder: Instances in parent group-v244700. [ 840.850116] env[63531]: DEBUG oslo.service.loopingcall [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 840.850332] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 840.850555] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-df7ff548-b4df-46e9-90f7-47d50b0568c3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.876321] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 840.876321] env[63531]: value = "task-1117906" [ 840.876321] env[63531]: _type = "Task" [ 840.876321] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.884038] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117906, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.116396] env[63531]: DEBUG oslo_vmware.api [None req-13b8019c-692f-4097-a7ee-b3251d4a991c tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117901, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.177495] env[63531]: DEBUG nova.scheduler.client.report [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 841.214475] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a81f1a0b-6718-4880-9962-0449389e3290 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Acquiring lock "e0074e2e-e8bd-4bcb-9ad9-271a10e079e4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.215021] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a81f1a0b-6718-4880-9962-0449389e3290 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Lock "e0074e2e-e8bd-4bcb-9ad9-271a10e079e4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.215319] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a81f1a0b-6718-4880-9962-0449389e3290 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Acquiring lock "e0074e2e-e8bd-4bcb-9ad9-271a10e079e4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.215546] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a81f1a0b-6718-4880-9962-0449389e3290 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Lock "e0074e2e-e8bd-4bcb-9ad9-271a10e079e4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.215769] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a81f1a0b-6718-4880-9962-0449389e3290 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Lock "e0074e2e-e8bd-4bcb-9ad9-271a10e079e4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.218090] env[63531]: INFO nova.compute.manager [None req-a81f1a0b-6718-4880-9962-0449389e3290 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Terminating instance [ 841.220017] env[63531]: DEBUG nova.compute.manager [None req-a81f1a0b-6718-4880-9962-0449389e3290 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 841.220232] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a81f1a0b-6718-4880-9962-0449389e3290 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 841.221275] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0fe7952-0808-4380-87ba-10a37a00bc3d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.230311] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-a81f1a0b-6718-4880-9962-0449389e3290 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 841.231126] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-68f3dadb-a392-417d-9939-b22a8e0d6ce7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.238208] env[63531]: DEBUG oslo_vmware.api [None req-a81f1a0b-6718-4880-9962-0449389e3290 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Waiting for the task: (returnval){ [ 841.238208] env[63531]: value = "task-1117907" [ 841.238208] env[63531]: _type = "Task" [ 841.238208] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.248964] env[63531]: DEBUG oslo_vmware.api [None req-a81f1a0b-6718-4880-9962-0449389e3290 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117907, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.276030] env[63531]: DEBUG nova.compute.manager [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 3eeaa6f9-60d9-43fe-a438-fbf2ceececae] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 841.301157] env[63531]: DEBUG nova.virt.hardware [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 841.301463] env[63531]: DEBUG nova.virt.hardware [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 841.301667] env[63531]: DEBUG nova.virt.hardware [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 841.301885] env[63531]: DEBUG nova.virt.hardware [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 841.302083] env[63531]: DEBUG nova.virt.hardware [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 841.302284] env[63531]: DEBUG nova.virt.hardware [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 841.302566] env[63531]: DEBUG nova.virt.hardware [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 841.302771] env[63531]: DEBUG nova.virt.hardware [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 841.302975] env[63531]: DEBUG nova.virt.hardware [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 841.303203] env[63531]: DEBUG nova.virt.hardware [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 841.303402] env[63531]: DEBUG nova.virt.hardware [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 841.304213] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc4e33ca-44ca-4722-950f-132a25339f93 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.309920] env[63531]: DEBUG oslo_vmware.api [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Task: {'id': task-1117902, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.315017] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85176ed1-5468-4ca3-9aa1-a30ed64b7459 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.330168] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 3eeaa6f9-60d9-43fe-a438-fbf2ceececae] Instance VIF info [] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 841.337075] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Creating folder: Project (3c08def9461647bb95c63055f258d46e). Parent ref: group-v244585. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 841.337075] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-09cedb1c-36e7-485a-a672-ec6873ee1826 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.350024] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Created folder: Project (3c08def9461647bb95c63055f258d46e) in parent group-v244585. [ 841.350245] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Creating folder: Instances. Parent ref: group-v244703. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 841.351249] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b85a90a5-0297-4ea6-80bc-be7806174c9a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.360157] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Created folder: Instances in parent group-v244703. [ 841.360548] env[63531]: DEBUG oslo.service.loopingcall [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 841.361070] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3eeaa6f9-60d9-43fe-a438-fbf2ceececae] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 841.361371] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-97b22759-b236-4d84-8553-30aa3684d6b1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.382241] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 841.382241] env[63531]: value = "task-1117910" [ 841.382241] env[63531]: _type = "Task" [ 841.382241] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.390032] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117906, 'name': CreateVM_Task} progress is 15%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.395098] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117910, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.459887] env[63531]: DEBUG nova.compute.manager [req-8f65eac3-262b-4115-acac-695c93add2e0 req-695adbae-77ca-4333-9289-5807cc9b4aa6 service nova] [instance: 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d] Received event network-changed-f7815862-8080-412a-8d54-7bad36f1413b {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 841.460275] env[63531]: DEBUG nova.compute.manager [req-8f65eac3-262b-4115-acac-695c93add2e0 req-695adbae-77ca-4333-9289-5807cc9b4aa6 service nova] [instance: 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d] Refreshing instance network info cache due to event network-changed-f7815862-8080-412a-8d54-7bad36f1413b. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 841.460550] env[63531]: DEBUG oslo_concurrency.lockutils [req-8f65eac3-262b-4115-acac-695c93add2e0 req-695adbae-77ca-4333-9289-5807cc9b4aa6 service nova] Acquiring lock "refresh_cache-56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.460747] env[63531]: DEBUG oslo_concurrency.lockutils [req-8f65eac3-262b-4115-acac-695c93add2e0 req-695adbae-77ca-4333-9289-5807cc9b4aa6 service nova] Acquired lock "refresh_cache-56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.460939] env[63531]: DEBUG nova.network.neutron [req-8f65eac3-262b-4115-acac-695c93add2e0 req-695adbae-77ca-4333-9289-5807cc9b4aa6 service nova] [instance: 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d] Refreshing network info cache for port f7815862-8080-412a-8d54-7bad36f1413b {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 841.615920] env[63531]: DEBUG oslo_vmware.api [None req-13b8019c-692f-4097-a7ee-b3251d4a991c tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117901, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.683072] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.427s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.683475] env[63531]: DEBUG nova.compute.manager [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 841.686306] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.003s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.688230] env[63531]: INFO nova.compute.claims [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 841.748943] env[63531]: DEBUG oslo_vmware.api [None req-a81f1a0b-6718-4880-9962-0449389e3290 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117907, 'name': PowerOffVM_Task, 'duration_secs': 0.190617} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.749234] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-a81f1a0b-6718-4880-9962-0449389e3290 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 841.749408] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a81f1a0b-6718-4880-9962-0449389e3290 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 841.749680] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-80006256-80e9-46d6-bdf6-be7a3c23808f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.807444] env[63531]: DEBUG oslo_vmware.api [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Task: {'id': task-1117902, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.844222] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a81f1a0b-6718-4880-9962-0449389e3290 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 841.844454] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a81f1a0b-6718-4880-9962-0449389e3290 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Deleting contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 841.845176] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-a81f1a0b-6718-4880-9962-0449389e3290 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Deleting the datastore file [datastore2] e0074e2e-e8bd-4bcb-9ad9-271a10e079e4 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 841.845176] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3bdddd82-ca49-41d0-a075-1bef7198ff14 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.853987] env[63531]: DEBUG oslo_vmware.api [None req-a81f1a0b-6718-4880-9962-0449389e3290 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Waiting for the task: (returnval){ [ 841.853987] env[63531]: value = "task-1117912" [ 841.853987] env[63531]: _type = "Task" [ 841.853987] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.866395] env[63531]: DEBUG oslo_vmware.api [None req-a81f1a0b-6718-4880-9962-0449389e3290 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117912, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.886469] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117906, 'name': CreateVM_Task} progress is 25%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.895639] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117910, 'name': CreateVM_Task} progress is 25%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.116396] env[63531]: DEBUG oslo_vmware.api [None req-13b8019c-692f-4097-a7ee-b3251d4a991c tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117901, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.196044] env[63531]: DEBUG nova.compute.utils [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 842.197481] env[63531]: DEBUG nova.compute.manager [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Not allocating networking since 'none' was specified. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 842.202769] env[63531]: DEBUG nova.network.neutron [req-8f65eac3-262b-4115-acac-695c93add2e0 req-695adbae-77ca-4333-9289-5807cc9b4aa6 service nova] [instance: 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d] Updated VIF entry in instance network info cache for port f7815862-8080-412a-8d54-7bad36f1413b. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 842.203204] env[63531]: DEBUG nova.network.neutron [req-8f65eac3-262b-4115-acac-695c93add2e0 req-695adbae-77ca-4333-9289-5807cc9b4aa6 service nova] [instance: 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d] Updating instance_info_cache with network_info: [{"id": "f7815862-8080-412a-8d54-7bad36f1413b", "address": "fa:16:3e:fd:cf:42", "network": {"id": "b47d3a83-c039-4eb2-b9b1-50898df74a35", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-465774433-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3e47ec69a0304ada957bbbcbf881460e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "da0e5087-d65b-416f-90fe-beaa9c534ad3", "external-id": "nsx-vlan-transportzone-522", "segmentation_id": 522, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf7815862-80", "ovs_interfaceid": "f7815862-8080-412a-8d54-7bad36f1413b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.305046] env[63531]: DEBUG oslo_vmware.api [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Task: {'id': task-1117902, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.791951} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.305046] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 45b05c54-6865-47a1-ba93-90ad3e1ba07e/45b05c54-6865-47a1-ba93-90ad3e1ba07e.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 842.305262] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 842.305589] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1f594e1d-4a44-456b-908d-c297c0c0a785 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.312712] env[63531]: DEBUG oslo_vmware.api [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Waiting for the task: (returnval){ [ 842.312712] env[63531]: value = "task-1117913" [ 842.312712] env[63531]: _type = "Task" [ 842.312712] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.320542] env[63531]: DEBUG oslo_vmware.api [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Task: {'id': task-1117913, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.364122] env[63531]: DEBUG oslo_vmware.api [None req-a81f1a0b-6718-4880-9962-0449389e3290 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1117912, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.292514} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.364418] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-a81f1a0b-6718-4880-9962-0449389e3290 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 842.364656] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a81f1a0b-6718-4880-9962-0449389e3290 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Deleted contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 842.364848] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a81f1a0b-6718-4880-9962-0449389e3290 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 842.364997] env[63531]: INFO nova.compute.manager [None req-a81f1a0b-6718-4880-9962-0449389e3290 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Took 1.14 seconds to destroy the instance on the hypervisor. [ 842.365280] env[63531]: DEBUG oslo.service.loopingcall [None req-a81f1a0b-6718-4880-9962-0449389e3290 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 842.365587] env[63531]: DEBUG nova.compute.manager [-] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 842.365587] env[63531]: DEBUG nova.network.neutron [-] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 842.387505] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117906, 'name': CreateVM_Task, 'duration_secs': 1.214495} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.390458] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 842.391202] env[63531]: DEBUG oslo_concurrency.lockutils [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.391368] env[63531]: DEBUG oslo_concurrency.lockutils [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.391693] env[63531]: DEBUG oslo_concurrency.lockutils [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 842.392384] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-52fa55b0-af56-48a1-9982-a967c7e33e57 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.398768] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117910, 'name': CreateVM_Task, 'duration_secs': 0.689985} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.399401] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3eeaa6f9-60d9-43fe-a438-fbf2ceececae] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 842.399782] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.401779] env[63531]: DEBUG oslo_vmware.api [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Waiting for the task: (returnval){ [ 842.401779] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]526cb593-d5f0-9f2c-8fe6-0e30b390c5b3" [ 842.401779] env[63531]: _type = "Task" [ 842.401779] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.409570] env[63531]: DEBUG oslo_vmware.api [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]526cb593-d5f0-9f2c-8fe6-0e30b390c5b3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.617013] env[63531]: DEBUG oslo_vmware.api [None req-13b8019c-692f-4097-a7ee-b3251d4a991c tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117901, 'name': RemoveSnapshot_Task, 'duration_secs': 2.400708} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.618672] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-13b8019c-692f-4097-a7ee-b3251d4a991c tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Deleted Snapshot of the VM instance {{(pid=63531) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 842.618916] env[63531]: INFO nova.compute.manager [None req-13b8019c-692f-4097-a7ee-b3251d4a991c tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Took 15.84 seconds to snapshot the instance on the hypervisor. [ 842.702297] env[63531]: DEBUG nova.compute.manager [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 842.705348] env[63531]: DEBUG oslo_concurrency.lockutils [req-8f65eac3-262b-4115-acac-695c93add2e0 req-695adbae-77ca-4333-9289-5807cc9b4aa6 service nova] Releasing lock "refresh_cache-56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.724110] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f18d56a-00b5-41f4-ab3b-5e091e729fdb tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Volume attach. Driver type: vmdk {{(pid=63531) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 842.724385] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f18d56a-00b5-41f4-ab3b-5e091e729fdb tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244696', 'volume_id': '02e72028-7ac8-4eab-88b0-9de83339ce64', 'name': 'volume-02e72028-7ac8-4eab-88b0-9de83339ce64', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '48043b5d-c0d3-4e8a-8d28-5a408d800e11', 'attached_at': '', 'detached_at': '', 'volume_id': '02e72028-7ac8-4eab-88b0-9de83339ce64', 'serial': '02e72028-7ac8-4eab-88b0-9de83339ce64'} {{(pid=63531) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 842.725264] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dfdf1fc-23f3-40e9-b088-db18ae02466a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.747626] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-980988cf-8098-4452-a816-c8086b15951e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.774049] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f18d56a-00b5-41f4-ab3b-5e091e729fdb tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Reconfiguring VM instance instance-00000037 to attach disk [datastore1] volume-02e72028-7ac8-4eab-88b0-9de83339ce64/volume-02e72028-7ac8-4eab-88b0-9de83339ce64.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 842.776094] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ecb85fbd-7fec-4158-abb3-1db400140e89 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.795448] env[63531]: DEBUG oslo_vmware.api [None req-8f18d56a-00b5-41f4-ab3b-5e091e729fdb tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for the task: (returnval){ [ 842.795448] env[63531]: value = "task-1117914" [ 842.795448] env[63531]: _type = "Task" [ 842.795448] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.806191] env[63531]: DEBUG oslo_vmware.api [None req-8f18d56a-00b5-41f4-ab3b-5e091e729fdb tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1117914, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.821546] env[63531]: DEBUG oslo_vmware.api [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Task: {'id': task-1117913, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074232} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.821824] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 842.822599] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6de0f3bf-5dde-43c8-8a24-c8700c31fdb5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.848330] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] Reconfiguring VM instance instance-0000003f to attach disk [datastore1] 45b05c54-6865-47a1-ba93-90ad3e1ba07e/45b05c54-6865-47a1-ba93-90ad3e1ba07e.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 842.851420] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1e33b63d-1d6c-478c-8ff0-df10b64d29ce {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.875325] env[63531]: DEBUG oslo_vmware.api [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Waiting for the task: (returnval){ [ 842.875325] env[63531]: value = "task-1117915" [ 842.875325] env[63531]: _type = "Task" [ 842.875325] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.881915] env[63531]: DEBUG oslo_vmware.api [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Task: {'id': task-1117915, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.917582] env[63531]: DEBUG oslo_vmware.api [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]526cb593-d5f0-9f2c-8fe6-0e30b390c5b3, 'name': SearchDatastore_Task, 'duration_secs': 0.010271} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.918210] env[63531]: DEBUG oslo_concurrency.lockutils [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.918210] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] [instance: 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 842.918488] env[63531]: DEBUG oslo_concurrency.lockutils [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.918655] env[63531]: DEBUG oslo_concurrency.lockutils [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.918852] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 842.919187] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.919529] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 842.919759] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-85c58d88-fdd2-4fb4-a3fd-de8833f3b0ba {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.922108] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1375d7fb-9126-48a6-acdf-358cad0962b9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.929849] env[63531]: DEBUG oslo_vmware.api [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Waiting for the task: (returnval){ [ 842.929849] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52fe6538-0d2e-d691-d88b-7adc8d75ff78" [ 842.929849] env[63531]: _type = "Task" [ 842.929849] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.934645] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 842.934926] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 842.936172] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-595ea266-4d5c-4c4b-9fff-114100c4649b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.944315] env[63531]: DEBUG oslo_vmware.api [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52fe6538-0d2e-d691-d88b-7adc8d75ff78, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.949134] env[63531]: DEBUG oslo_vmware.api [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Waiting for the task: (returnval){ [ 842.949134] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52e485be-14f9-99b9-ca44-8e3959859b47" [ 842.949134] env[63531]: _type = "Task" [ 842.949134] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.955649] env[63531]: DEBUG oslo_vmware.api [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52e485be-14f9-99b9-ca44-8e3959859b47, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.121946] env[63531]: DEBUG nova.network.neutron [-] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 843.161293] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-378ba059-f58f-45f3-9130-c89f860d55dc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.168875] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-679d12f5-8ae3-459b-a5c5-5e69114fed19 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.200390] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0081e69c-6d94-4204-8213-e6265aea6c6d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.211091] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e05a5fc-cac8-4294-9022-11c29f93c1dd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.224973] env[63531]: DEBUG nova.compute.provider_tree [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 843.305346] env[63531]: DEBUG oslo_vmware.api [None req-8f18d56a-00b5-41f4-ab3b-5e091e729fdb tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1117914, 'name': ReconfigVM_Task, 'duration_secs': 0.372415} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.305666] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f18d56a-00b5-41f4-ab3b-5e091e729fdb tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Reconfigured VM instance instance-00000037 to attach disk [datastore1] volume-02e72028-7ac8-4eab-88b0-9de83339ce64/volume-02e72028-7ac8-4eab-88b0-9de83339ce64.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 843.310425] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0f3d3f97-c95e-47ae-ada3-ef91905cc953 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.326029] env[63531]: DEBUG oslo_vmware.api [None req-8f18d56a-00b5-41f4-ab3b-5e091e729fdb tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for the task: (returnval){ [ 843.326029] env[63531]: value = "task-1117916" [ 843.326029] env[63531]: _type = "Task" [ 843.326029] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.334530] env[63531]: DEBUG oslo_vmware.api [None req-8f18d56a-00b5-41f4-ab3b-5e091e729fdb tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1117916, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.382822] env[63531]: DEBUG oslo_vmware.api [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Task: {'id': task-1117915, 'name': ReconfigVM_Task, 'duration_secs': 0.331534} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.383053] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] Reconfigured VM instance instance-0000003f to attach disk [datastore1] 45b05c54-6865-47a1-ba93-90ad3e1ba07e/45b05c54-6865-47a1-ba93-90ad3e1ba07e.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 843.384286] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-859171f4-c334-4033-a31e-80577cba2615 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.391870] env[63531]: DEBUG oslo_vmware.api [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Waiting for the task: (returnval){ [ 843.391870] env[63531]: value = "task-1117917" [ 843.391870] env[63531]: _type = "Task" [ 843.391870] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.404278] env[63531]: DEBUG oslo_vmware.api [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Task: {'id': task-1117917, 'name': Rename_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.440941] env[63531]: DEBUG oslo_vmware.api [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52fe6538-0d2e-d691-d88b-7adc8d75ff78, 'name': SearchDatastore_Task, 'duration_secs': 0.011299} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.441399] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.441798] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 3eeaa6f9-60d9-43fe-a438-fbf2ceececae] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 843.442196] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.458742] env[63531]: DEBUG oslo_vmware.api [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52e485be-14f9-99b9-ca44-8e3959859b47, 'name': SearchDatastore_Task, 'duration_secs': 0.009999} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.459630] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ef233071-9222-448f-87b6-6413508ef8ff {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.465596] env[63531]: DEBUG oslo_vmware.api [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Waiting for the task: (returnval){ [ 843.465596] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]528bed31-e45d-b39c-dccb-937204e77540" [ 843.465596] env[63531]: _type = "Task" [ 843.465596] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.475534] env[63531]: DEBUG oslo_vmware.api [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]528bed31-e45d-b39c-dccb-937204e77540, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.537676] env[63531]: DEBUG nova.compute.manager [req-8d41e2fa-7013-44ba-afec-25254c777660 req-34f0f8dd-7011-47be-a7b3-2f01baac748f service nova] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Received event network-vif-deleted-f1ef6f0e-1f62-4034-8edb-e8cc1bb98abd {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 843.626655] env[63531]: INFO nova.compute.manager [-] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Took 1.26 seconds to deallocate network for instance. [ 843.717327] env[63531]: DEBUG nova.compute.manager [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 843.727626] env[63531]: DEBUG nova.scheduler.client.report [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 843.738782] env[63531]: DEBUG nova.virt.hardware [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 843.739042] env[63531]: DEBUG nova.virt.hardware [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 843.739209] env[63531]: DEBUG nova.virt.hardware [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 843.739395] env[63531]: DEBUG nova.virt.hardware [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 843.739543] env[63531]: DEBUG nova.virt.hardware [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 843.739693] env[63531]: DEBUG nova.virt.hardware [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 843.739904] env[63531]: DEBUG nova.virt.hardware [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 843.740104] env[63531]: DEBUG nova.virt.hardware [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 843.740305] env[63531]: DEBUG nova.virt.hardware [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 843.740562] env[63531]: DEBUG nova.virt.hardware [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 843.740680] env[63531]: DEBUG nova.virt.hardware [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 843.741559] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-334017f7-ff68-4498-8292-6e6db9052454 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.750781] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28e6003a-3ed0-41c3-bdb9-a9ffbbdfea8a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.764859] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Instance VIF info [] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 843.770545] env[63531]: DEBUG oslo.service.loopingcall [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 843.770829] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 843.771062] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a4b0c2d6-101c-43a8-966d-15debbf31f37 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.787670] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 843.787670] env[63531]: value = "task-1117918" [ 843.787670] env[63531]: _type = "Task" [ 843.787670] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.796759] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117918, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.838709] env[63531]: DEBUG oslo_vmware.api [None req-8f18d56a-00b5-41f4-ab3b-5e091e729fdb tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1117916, 'name': ReconfigVM_Task, 'duration_secs': 0.154467} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.839097] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f18d56a-00b5-41f4-ab3b-5e091e729fdb tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244696', 'volume_id': '02e72028-7ac8-4eab-88b0-9de83339ce64', 'name': 'volume-02e72028-7ac8-4eab-88b0-9de83339ce64', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '48043b5d-c0d3-4e8a-8d28-5a408d800e11', 'attached_at': '', 'detached_at': '', 'volume_id': '02e72028-7ac8-4eab-88b0-9de83339ce64', 'serial': '02e72028-7ac8-4eab-88b0-9de83339ce64'} {{(pid=63531) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 843.903933] env[63531]: DEBUG oslo_vmware.api [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Task: {'id': task-1117917, 'name': Rename_Task, 'duration_secs': 0.145555} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.903933] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 843.904302] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7d673c57-ffd3-4bc7-88cc-790d1c60bdfb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.911021] env[63531]: DEBUG oslo_vmware.api [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Waiting for the task: (returnval){ [ 843.911021] env[63531]: value = "task-1117919" [ 843.911021] env[63531]: _type = "Task" [ 843.911021] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.923071] env[63531]: DEBUG oslo_vmware.api [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Task: {'id': task-1117919, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.975846] env[63531]: DEBUG oslo_vmware.api [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]528bed31-e45d-b39c-dccb-937204e77540, 'name': SearchDatastore_Task, 'duration_secs': 0.01132} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.976186] env[63531]: DEBUG oslo_concurrency.lockutils [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.976498] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d/56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 843.976791] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.977010] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 843.977262] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6a886938-3f62-432b-9e63-53c5e98deca8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.979599] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f482f21d-c838-4266-b1a5-cd321ba7e115 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.985924] env[63531]: DEBUG oslo_vmware.api [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Waiting for the task: (returnval){ [ 843.985924] env[63531]: value = "task-1117920" [ 843.985924] env[63531]: _type = "Task" [ 843.985924] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.991252] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 843.991252] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 843.992453] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9a0287cb-d3cd-403c-b057-545bdb6e5935 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.997867] env[63531]: DEBUG oslo_vmware.api [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Task: {'id': task-1117920, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.001087] env[63531]: DEBUG oslo_vmware.api [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Waiting for the task: (returnval){ [ 844.001087] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52b2d94f-a364-9cc7-8d41-f134e60cb9ca" [ 844.001087] env[63531]: _type = "Task" [ 844.001087] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.009853] env[63531]: DEBUG oslo_vmware.api [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52b2d94f-a364-9cc7-8d41-f134e60cb9ca, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.133823] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a81f1a0b-6718-4880-9962-0449389e3290 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.234244] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.548s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.234835] env[63531]: DEBUG nova.compute.manager [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 844.237750] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.461s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.239311] env[63531]: INFO nova.compute.claims [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 844.298609] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117918, 'name': CreateVM_Task, 'duration_secs': 0.357264} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.298818] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 844.299330] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.299502] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.299870] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 844.300180] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-84322f6d-e877-43ff-84ed-f9c8949e00ed {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.305716] env[63531]: DEBUG oslo_vmware.api [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Waiting for the task: (returnval){ [ 844.305716] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52573fa9-d812-74a1-adaf-c9d2a698918b" [ 844.305716] env[63531]: _type = "Task" [ 844.305716] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.314113] env[63531]: DEBUG oslo_vmware.api [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52573fa9-d812-74a1-adaf-c9d2a698918b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.424939] env[63531]: DEBUG oslo_vmware.api [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Task: {'id': task-1117919, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.497556] env[63531]: DEBUG oslo_vmware.api [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Task: {'id': task-1117920, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.511983] env[63531]: DEBUG oslo_vmware.api [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52b2d94f-a364-9cc7-8d41-f134e60cb9ca, 'name': SearchDatastore_Task, 'duration_secs': 0.015231} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.512772] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-59771c7f-f1ab-4ef2-a06e-06b5f34af275 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.519408] env[63531]: DEBUG oslo_vmware.api [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Waiting for the task: (returnval){ [ 844.519408] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5219d853-14fd-5cda-94dd-d379a20f1e73" [ 844.519408] env[63531]: _type = "Task" [ 844.519408] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.528326] env[63531]: DEBUG oslo_vmware.api [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5219d853-14fd-5cda-94dd-d379a20f1e73, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.746438] env[63531]: DEBUG nova.compute.utils [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 844.747865] env[63531]: DEBUG nova.compute.manager [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 844.748175] env[63531]: DEBUG nova.network.neutron [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 844.821557] env[63531]: DEBUG oslo_vmware.api [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52573fa9-d812-74a1-adaf-c9d2a698918b, 'name': SearchDatastore_Task, 'duration_secs': 0.05584} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.822127] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.822417] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 844.822786] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.832124] env[63531]: DEBUG nova.policy [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '049cb89a769f41bd91d0e62784635cbd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c2e99349303e413b9f7896d449cb7dff', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 844.883221] env[63531]: DEBUG nova.objects.instance [None req-8f18d56a-00b5-41f4-ab3b-5e091e729fdb tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lazy-loading 'flavor' on Instance uuid 48043b5d-c0d3-4e8a-8d28-5a408d800e11 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 844.922535] env[63531]: DEBUG oslo_vmware.api [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Task: {'id': task-1117919, 'name': PowerOnVM_Task, 'duration_secs': 0.658212} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.923913] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 844.923913] env[63531]: INFO nova.compute.manager [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] Took 8.80 seconds to spawn the instance on the hypervisor. [ 844.923913] env[63531]: DEBUG nova.compute.manager [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 844.924268] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9333238d-5849-4cb5-aff3-5e133bac719f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.998927] env[63531]: DEBUG oslo_vmware.api [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Task: {'id': task-1117920, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.593518} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.999271] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d/56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 844.999500] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] [instance: 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 844.999804] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-24d18beb-a4cb-4b1a-af9a-f6f9bf136554 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.006868] env[63531]: DEBUG oslo_vmware.api [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Waiting for the task: (returnval){ [ 845.006868] env[63531]: value = "task-1117921" [ 845.006868] env[63531]: _type = "Task" [ 845.006868] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.015608] env[63531]: DEBUG oslo_vmware.api [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Task: {'id': task-1117921, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.034020] env[63531]: DEBUG oslo_vmware.api [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5219d853-14fd-5cda-94dd-d379a20f1e73, 'name': SearchDatastore_Task, 'duration_secs': 0.039717} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.036066] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.036066] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 3eeaa6f9-60d9-43fe-a438-fbf2ceececae/3eeaa6f9-60d9-43fe-a438-fbf2ceececae.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 845.036066] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.036066] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 845.036066] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-752012b3-bdc5-4bdb-9a7e-b22c9caa0d48 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.040166] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5452b272-ce13-4e89-ad1a-be331ef71110 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.046380] env[63531]: DEBUG oslo_vmware.api [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Waiting for the task: (returnval){ [ 845.046380] env[63531]: value = "task-1117922" [ 845.046380] env[63531]: _type = "Task" [ 845.046380] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.055584] env[63531]: DEBUG oslo_vmware.api [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': task-1117922, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.057963] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 845.058620] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 845.059562] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0c38a082-f4b9-47b2-8fd5-141cd5535aa7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.065327] env[63531]: DEBUG oslo_vmware.api [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Waiting for the task: (returnval){ [ 845.065327] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52689cbd-c369-bd29-29df-5584dd34c506" [ 845.065327] env[63531]: _type = "Task" [ 845.065327] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.076095] env[63531]: DEBUG oslo_vmware.api [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52689cbd-c369-bd29-29df-5584dd34c506, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.178987] env[63531]: DEBUG nova.network.neutron [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Successfully created port: 13365dfb-f81c-4e56-bf30-0a09dec4f561 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 845.252587] env[63531]: DEBUG nova.compute.manager [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 845.389226] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8f18d56a-00b5-41f4-ab3b-5e091e729fdb tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "48043b5d-c0d3-4e8a-8d28-5a408d800e11" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.311s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.454777] env[63531]: INFO nova.compute.manager [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] Took 52.11 seconds to build instance. [ 845.520721] env[63531]: DEBUG oslo_vmware.api [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Task: {'id': task-1117921, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.085692} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.521029] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] [instance: 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 845.521874] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd970794-27ec-4cd8-a04d-a24a03cd78cd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.547711] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] [instance: 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d] Reconfiguring VM instance instance-00000040 to attach disk [datastore1] 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d/56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 845.550948] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6e7b8435-ac73-412a-93e8-75c4970f8ac3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.579960] env[63531]: DEBUG oslo_vmware.api [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': task-1117922, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.497235} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.586957] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 3eeaa6f9-60d9-43fe-a438-fbf2ceececae/3eeaa6f9-60d9-43fe-a438-fbf2ceececae.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 845.587288] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 3eeaa6f9-60d9-43fe-a438-fbf2ceececae] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 845.587572] env[63531]: DEBUG oslo_vmware.api [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52689cbd-c369-bd29-29df-5584dd34c506, 'name': SearchDatastore_Task, 'duration_secs': 0.013838} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.587858] env[63531]: DEBUG oslo_vmware.api [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Waiting for the task: (returnval){ [ 845.587858] env[63531]: value = "task-1117923" [ 845.587858] env[63531]: _type = "Task" [ 845.587858] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.588897] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cacc03cb-1648-484b-aaf2-ea15e131f9f6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.591397] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bc0fd225-54c7-4381-9f3a-9f692ed0e549 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.599649] env[63531]: DEBUG oslo_vmware.api [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Waiting for the task: (returnval){ [ 845.599649] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52a2d5f0-fe75-bc4e-c37a-7c2420f05ffc" [ 845.599649] env[63531]: _type = "Task" [ 845.599649] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.603233] env[63531]: DEBUG oslo_vmware.api [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Task: {'id': task-1117923, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.610019] env[63531]: DEBUG oslo_vmware.api [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Waiting for the task: (returnval){ [ 845.610019] env[63531]: value = "task-1117924" [ 845.610019] env[63531]: _type = "Task" [ 845.610019] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.618483] env[63531]: DEBUG oslo_vmware.api [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52a2d5f0-fe75-bc4e-c37a-7c2420f05ffc, 'name': SearchDatastore_Task, 'duration_secs': 0.008931} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.623034] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.623034] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 09feb12a-da9f-4bb7-959d-84c5c7c58e99/09feb12a-da9f-4bb7-959d-84c5c7c58e99.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 845.623034] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-31a1a3b9-6478-4ab5-8cd8-406d5c5afef8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.626905] env[63531]: DEBUG oslo_vmware.api [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': task-1117924, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.631052] env[63531]: DEBUG oslo_vmware.api [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Waiting for the task: (returnval){ [ 845.631052] env[63531]: value = "task-1117925" [ 845.631052] env[63531]: _type = "Task" [ 845.631052] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.642506] env[63531]: DEBUG oslo_vmware.api [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': task-1117925, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.727311] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a2338418-703e-4117-a6e8-2c9a448381e7 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Acquiring lock "64fb5831-0789-482b-bb98-67b29868c4c9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.727311] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a2338418-703e-4117-a6e8-2c9a448381e7 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Lock "64fb5831-0789-482b-bb98-67b29868c4c9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.728331] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a2338418-703e-4117-a6e8-2c9a448381e7 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Acquiring lock "64fb5831-0789-482b-bb98-67b29868c4c9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.728331] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a2338418-703e-4117-a6e8-2c9a448381e7 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Lock "64fb5831-0789-482b-bb98-67b29868c4c9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.728331] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a2338418-703e-4117-a6e8-2c9a448381e7 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Lock "64fb5831-0789-482b-bb98-67b29868c4c9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.730159] env[63531]: INFO nova.compute.manager [None req-a2338418-703e-4117-a6e8-2c9a448381e7 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Terminating instance [ 845.737388] env[63531]: DEBUG nova.compute.manager [None req-a2338418-703e-4117-a6e8-2c9a448381e7 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 845.737388] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a2338418-703e-4117-a6e8-2c9a448381e7 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 845.737388] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32bb9d6c-697c-47ef-99e8-43b0cc60b2ef {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.743901] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2338418-703e-4117-a6e8-2c9a448381e7 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 845.744172] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5b2fa1af-f3ef-4579-a1cc-3ffa1006dd5c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.749899] env[63531]: DEBUG oslo_vmware.api [None req-a2338418-703e-4117-a6e8-2c9a448381e7 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Waiting for the task: (returnval){ [ 845.749899] env[63531]: value = "task-1117926" [ 845.749899] env[63531]: _type = "Task" [ 845.749899] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.763747] env[63531]: DEBUG oslo_vmware.api [None req-a2338418-703e-4117-a6e8-2c9a448381e7 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117926, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.776309] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47fc57c7-020d-424b-9269-f288a7234134 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.784427] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a989fcd2-4ebb-498e-bb8f-ea1394f73ff3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.825765] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7b3c683-20e7-4ccf-87d0-7e4f81641b3c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.839662] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30fec520-605e-4b13-85e8-89bb80f24937 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.856983] env[63531]: DEBUG nova.compute.provider_tree [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 845.958738] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e455dc88-77e5-4087-bec9-27ff9898513d tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Lock "45b05c54-6865-47a1-ba93-90ad3e1ba07e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 54.225s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.102358] env[63531]: DEBUG oslo_vmware.api [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Task: {'id': task-1117923, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.119670] env[63531]: DEBUG oslo_vmware.api [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': task-1117924, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.141600] env[63531]: DEBUG oslo_vmware.api [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': task-1117925, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.454335} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.142411] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 09feb12a-da9f-4bb7-959d-84c5c7c58e99/09feb12a-da9f-4bb7-959d-84c5c7c58e99.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 846.142411] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 846.142411] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3acd69ca-853a-4210-ad30-7a97d4eb924c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.149175] env[63531]: DEBUG oslo_vmware.api [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Waiting for the task: (returnval){ [ 846.149175] env[63531]: value = "task-1117927" [ 846.149175] env[63531]: _type = "Task" [ 846.149175] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.157816] env[63531]: DEBUG oslo_vmware.api [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': task-1117927, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.260934] env[63531]: DEBUG oslo_vmware.api [None req-a2338418-703e-4117-a6e8-2c9a448381e7 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117926, 'name': PowerOffVM_Task, 'duration_secs': 0.399973} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.261346] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2338418-703e-4117-a6e8-2c9a448381e7 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 846.261530] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a2338418-703e-4117-a6e8-2c9a448381e7 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 846.263333] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-42671144-d956-4b8b-baaa-fb53bc661883 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.265701] env[63531]: DEBUG nova.compute.manager [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 846.291385] env[63531]: DEBUG oslo_vmware.rw_handles [None req-92007411-3bac-410d-a9c6-c0b9a37679fa tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f3461f-e2c4-1251-e983-2cc3133ddf8a/disk-0.vmdk. {{(pid=63531) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 846.293569] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9577837-1e5e-4e07-ad32-ab7dff351830 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.300371] env[63531]: DEBUG oslo_vmware.rw_handles [None req-92007411-3bac-410d-a9c6-c0b9a37679fa tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f3461f-e2c4-1251-e983-2cc3133ddf8a/disk-0.vmdk is in state: ready. {{(pid=63531) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 846.300635] env[63531]: ERROR oslo_vmware.rw_handles [None req-92007411-3bac-410d-a9c6-c0b9a37679fa tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f3461f-e2c4-1251-e983-2cc3133ddf8a/disk-0.vmdk due to incomplete transfer. [ 846.302807] env[63531]: DEBUG nova.virt.hardware [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 846.303061] env[63531]: DEBUG nova.virt.hardware [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 846.303249] env[63531]: DEBUG nova.virt.hardware [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 846.303463] env[63531]: DEBUG nova.virt.hardware [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 846.303632] env[63531]: DEBUG nova.virt.hardware [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 846.303800] env[63531]: DEBUG nova.virt.hardware [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 846.304044] env[63531]: DEBUG nova.virt.hardware [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 846.304232] env[63531]: DEBUG nova.virt.hardware [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 846.304431] env[63531]: DEBUG nova.virt.hardware [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 846.304606] env[63531]: DEBUG nova.virt.hardware [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 846.304817] env[63531]: DEBUG nova.virt.hardware [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 846.305178] env[63531]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-cc51e695-6e5a-4f03-b05a-9ba6d15bca0b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.307507] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-638b306b-aac3-425e-8009-a98c7d9777b4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.317037] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d851a1ae-369f-413f-9da5-b83aea3d35db {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.320965] env[63531]: DEBUG oslo_vmware.rw_handles [None req-92007411-3bac-410d-a9c6-c0b9a37679fa tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f3461f-e2c4-1251-e983-2cc3133ddf8a/disk-0.vmdk. {{(pid=63531) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 846.321177] env[63531]: DEBUG nova.virt.vmwareapi.images [None req-92007411-3bac-410d-a9c6-c0b9a37679fa tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Uploaded image 4f765107-1431-4888-adf7-e8b9dbcda251 to the Glance image server {{(pid=63531) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 846.323045] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-92007411-3bac-410d-a9c6-c0b9a37679fa tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Destroying the VM {{(pid=63531) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 846.323293] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-fa138940-555e-4033-8fca-4cc1d8359731 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.336125] env[63531]: DEBUG oslo_vmware.api [None req-92007411-3bac-410d-a9c6-c0b9a37679fa tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 846.336125] env[63531]: value = "task-1117929" [ 846.336125] env[63531]: _type = "Task" [ 846.336125] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.344049] env[63531]: DEBUG oslo_vmware.api [None req-92007411-3bac-410d-a9c6-c0b9a37679fa tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1117929, 'name': Destroy_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.360308] env[63531]: DEBUG nova.scheduler.client.report [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 846.370604] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a2338418-703e-4117-a6e8-2c9a448381e7 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 846.370873] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a2338418-703e-4117-a6e8-2c9a448381e7 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 846.371089] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-a2338418-703e-4117-a6e8-2c9a448381e7 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Deleting the datastore file [datastore1] 64fb5831-0789-482b-bb98-67b29868c4c9 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 846.371403] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7febc397-f08b-414f-aea3-6b000f044742 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.377902] env[63531]: DEBUG oslo_vmware.api [None req-a2338418-703e-4117-a6e8-2c9a448381e7 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Waiting for the task: (returnval){ [ 846.377902] env[63531]: value = "task-1117930" [ 846.377902] env[63531]: _type = "Task" [ 846.377902] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.386988] env[63531]: DEBUG oslo_vmware.api [None req-a2338418-703e-4117-a6e8-2c9a448381e7 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117930, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.601860] env[63531]: DEBUG oslo_vmware.api [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Task: {'id': task-1117923, 'name': ReconfigVM_Task, 'duration_secs': 0.830253} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.602233] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] [instance: 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d] Reconfigured VM instance instance-00000040 to attach disk [datastore1] 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d/56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 846.602966] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-04fcc8cf-ab54-4b08-8306-67b4adcbaa59 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.609056] env[63531]: DEBUG oslo_vmware.api [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Waiting for the task: (returnval){ [ 846.609056] env[63531]: value = "task-1117931" [ 846.609056] env[63531]: _type = "Task" [ 846.609056] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.621913] env[63531]: DEBUG oslo_vmware.api [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Task: {'id': task-1117931, 'name': Rename_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.624745] env[63531]: DEBUG oslo_vmware.api [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': task-1117924, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.597181} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.625125] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 3eeaa6f9-60d9-43fe-a438-fbf2ceececae] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 846.626032] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38fce138-d1fe-4b14-80d6-4259be41d2c5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.649831] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 3eeaa6f9-60d9-43fe-a438-fbf2ceececae] Reconfiguring VM instance instance-00000041 to attach disk [datastore1] 3eeaa6f9-60d9-43fe-a438-fbf2ceececae/3eeaa6f9-60d9-43fe-a438-fbf2ceececae.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 846.651414] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-469119d1-ed8e-405e-a570-b1f55ba623f7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.678246] env[63531]: DEBUG oslo_vmware.api [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': task-1117927, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.085502} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.679683] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 846.680096] env[63531]: DEBUG oslo_vmware.api [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Waiting for the task: (returnval){ [ 846.680096] env[63531]: value = "task-1117932" [ 846.680096] env[63531]: _type = "Task" [ 846.680096] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.680910] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0a36d6b-50bb-4f53-9d98-afc6712dffb6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.705344] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Reconfiguring VM instance instance-00000042 to attach disk [datastore1] 09feb12a-da9f-4bb7-959d-84c5c7c58e99/09feb12a-da9f-4bb7-959d-84c5c7c58e99.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 846.710033] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ecbba6e0-326c-469b-bfcc-b70f1cf1e89e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.725163] env[63531]: DEBUG oslo_vmware.api [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': task-1117932, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.730754] env[63531]: DEBUG oslo_vmware.api [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Waiting for the task: (returnval){ [ 846.730754] env[63531]: value = "task-1117933" [ 846.730754] env[63531]: _type = "Task" [ 846.730754] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.740514] env[63531]: DEBUG oslo_vmware.api [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': task-1117933, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.846095] env[63531]: DEBUG oslo_vmware.api [None req-92007411-3bac-410d-a9c6-c0b9a37679fa tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1117929, 'name': Destroy_Task, 'duration_secs': 0.327283} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.846575] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-92007411-3bac-410d-a9c6-c0b9a37679fa tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Destroyed the VM [ 846.846882] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-92007411-3bac-410d-a9c6-c0b9a37679fa tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Deleting Snapshot of the VM instance {{(pid=63531) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 846.847398] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-4b148947-42fb-4bbf-8e01-33eee5a1b4df {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.853724] env[63531]: DEBUG oslo_vmware.api [None req-92007411-3bac-410d-a9c6-c0b9a37679fa tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 846.853724] env[63531]: value = "task-1117934" [ 846.853724] env[63531]: _type = "Task" [ 846.853724] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.862270] env[63531]: DEBUG oslo_vmware.api [None req-92007411-3bac-410d-a9c6-c0b9a37679fa tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1117934, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.869423] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.632s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.870015] env[63531]: DEBUG nova.compute.manager [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 846.872868] env[63531]: DEBUG oslo_concurrency.lockutils [None req-34c5a484-e724-4b75-a347-4f418835aae3 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.689s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.873147] env[63531]: DEBUG oslo_concurrency.lockutils [None req-34c5a484-e724-4b75-a347-4f418835aae3 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.876207] env[63531]: DEBUG oslo_concurrency.lockutils [None req-056f71aa-40aa-436b-8ae3-9a3a30a67adc tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.491s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.876791] env[63531]: DEBUG oslo_concurrency.lockutils [None req-056f71aa-40aa-436b-8ae3-9a3a30a67adc tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.878372] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.810s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.882021] env[63531]: INFO nova.compute.claims [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: db83adf4-6183-4330-b260-77d1f5daf899] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 846.892354] env[63531]: DEBUG oslo_vmware.api [None req-a2338418-703e-4117-a6e8-2c9a448381e7 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1117930, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.399068} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.893248] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-a2338418-703e-4117-a6e8-2c9a448381e7 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 846.893248] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a2338418-703e-4117-a6e8-2c9a448381e7 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 846.893248] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a2338418-703e-4117-a6e8-2c9a448381e7 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 846.894224] env[63531]: INFO nova.compute.manager [None req-a2338418-703e-4117-a6e8-2c9a448381e7 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Took 1.16 seconds to destroy the instance on the hypervisor. [ 846.894224] env[63531]: DEBUG oslo.service.loopingcall [None req-a2338418-703e-4117-a6e8-2c9a448381e7 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 846.894224] env[63531]: DEBUG nova.compute.manager [-] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 846.894529] env[63531]: DEBUG nova.network.neutron [-] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 846.913112] env[63531]: INFO nova.scheduler.client.report [None req-34c5a484-e724-4b75-a347-4f418835aae3 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Deleted allocations for instance 531e1852-0c67-4b4d-a0ca-749ac438e79a [ 846.914951] env[63531]: INFO nova.scheduler.client.report [None req-056f71aa-40aa-436b-8ae3-9a3a30a67adc tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Deleted allocations for instance caf0dee8-cb5a-402b-b533-22a590433f54 [ 846.926259] env[63531]: DEBUG nova.network.neutron [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Successfully updated port: 13365dfb-f81c-4e56-bf30-0a09dec4f561 {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 847.119377] env[63531]: DEBUG oslo_vmware.api [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Task: {'id': task-1117931, 'name': Rename_Task, 'duration_secs': 0.148458} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.119725] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] [instance: 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 847.119987] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7e0a7f8d-e85d-4d30-b0d7-995d94417476 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.126397] env[63531]: DEBUG oslo_vmware.api [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Waiting for the task: (returnval){ [ 847.126397] env[63531]: value = "task-1117935" [ 847.126397] env[63531]: _type = "Task" [ 847.126397] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.134149] env[63531]: DEBUG oslo_vmware.api [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Task: {'id': task-1117935, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.193877] env[63531]: DEBUG oslo_vmware.api [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': task-1117932, 'name': ReconfigVM_Task, 'duration_secs': 0.319979} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.194166] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 3eeaa6f9-60d9-43fe-a438-fbf2ceececae] Reconfigured VM instance instance-00000041 to attach disk [datastore1] 3eeaa6f9-60d9-43fe-a438-fbf2ceececae/3eeaa6f9-60d9-43fe-a438-fbf2ceececae.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 847.194859] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-562665e9-3dd4-44ea-a04c-1ea7f28eb1d7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.200538] env[63531]: DEBUG oslo_vmware.api [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Waiting for the task: (returnval){ [ 847.200538] env[63531]: value = "task-1117936" [ 847.200538] env[63531]: _type = "Task" [ 847.200538] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.211365] env[63531]: DEBUG oslo_vmware.api [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': task-1117936, 'name': Rename_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.233384] env[63531]: DEBUG nova.compute.manager [req-b8954be6-00fe-476f-9a6d-8dc7a954951f req-76b64762-370b-46a3-b5a5-2383e4552eff service nova] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Received event network-vif-plugged-13365dfb-f81c-4e56-bf30-0a09dec4f561 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 847.233609] env[63531]: DEBUG oslo_concurrency.lockutils [req-b8954be6-00fe-476f-9a6d-8dc7a954951f req-76b64762-370b-46a3-b5a5-2383e4552eff service nova] Acquiring lock "2b0e4285-310a-439c-bc50-000a7c5ef7f9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.233836] env[63531]: DEBUG oslo_concurrency.lockutils [req-b8954be6-00fe-476f-9a6d-8dc7a954951f req-76b64762-370b-46a3-b5a5-2383e4552eff service nova] Lock "2b0e4285-310a-439c-bc50-000a7c5ef7f9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.233998] env[63531]: DEBUG oslo_concurrency.lockutils [req-b8954be6-00fe-476f-9a6d-8dc7a954951f req-76b64762-370b-46a3-b5a5-2383e4552eff service nova] Lock "2b0e4285-310a-439c-bc50-000a7c5ef7f9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.235123] env[63531]: DEBUG nova.compute.manager [req-b8954be6-00fe-476f-9a6d-8dc7a954951f req-76b64762-370b-46a3-b5a5-2383e4552eff service nova] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] No waiting events found dispatching network-vif-plugged-13365dfb-f81c-4e56-bf30-0a09dec4f561 {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 847.235377] env[63531]: WARNING nova.compute.manager [req-b8954be6-00fe-476f-9a6d-8dc7a954951f req-76b64762-370b-46a3-b5a5-2383e4552eff service nova] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Received unexpected event network-vif-plugged-13365dfb-f81c-4e56-bf30-0a09dec4f561 for instance with vm_state building and task_state spawning. [ 847.246212] env[63531]: DEBUG oslo_vmware.api [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': task-1117933, 'name': ReconfigVM_Task, 'duration_secs': 0.321374} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.246521] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Reconfigured VM instance instance-00000042 to attach disk [datastore1] 09feb12a-da9f-4bb7-959d-84c5c7c58e99/09feb12a-da9f-4bb7-959d-84c5c7c58e99.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 847.247156] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b59c8393-10a4-410b-a6bb-460b57970004 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.254884] env[63531]: DEBUG oslo_vmware.api [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Waiting for the task: (returnval){ [ 847.254884] env[63531]: value = "task-1117937" [ 847.254884] env[63531]: _type = "Task" [ 847.254884] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.263327] env[63531]: DEBUG oslo_vmware.api [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': task-1117937, 'name': Rename_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.370693] env[63531]: DEBUG oslo_vmware.api [None req-92007411-3bac-410d-a9c6-c0b9a37679fa tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1117934, 'name': RemoveSnapshot_Task} progress is 26%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.386733] env[63531]: DEBUG nova.compute.utils [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 847.392496] env[63531]: DEBUG nova.compute.manager [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 847.392496] env[63531]: DEBUG nova.network.neutron [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 847.430205] env[63531]: DEBUG oslo_concurrency.lockutils [None req-34c5a484-e724-4b75-a347-4f418835aae3 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Lock "531e1852-0c67-4b4d-a0ca-749ac438e79a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.208s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.433386] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "refresh_cache-2b0e4285-310a-439c-bc50-000a7c5ef7f9" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.433764] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquired lock "refresh_cache-2b0e4285-310a-439c-bc50-000a7c5ef7f9" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.434356] env[63531]: DEBUG nova.network.neutron [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 847.438320] env[63531]: DEBUG oslo_concurrency.lockutils [None req-056f71aa-40aa-436b-8ae3-9a3a30a67adc tempest-SecurityGroupsTestJSON-300897749 tempest-SecurityGroupsTestJSON-300897749-project-member] Lock "caf0dee8-cb5a-402b-b533-22a590433f54" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.978s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.469770] env[63531]: DEBUG nova.policy [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1af2cb839598417b897695b83e867058', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8bebde0e281d485ebdd70a1a3b613632', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 847.637817] env[63531]: DEBUG oslo_vmware.api [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Task: {'id': task-1117935, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.711101] env[63531]: DEBUG oslo_vmware.api [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': task-1117936, 'name': Rename_Task, 'duration_secs': 0.183788} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.711417] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 3eeaa6f9-60d9-43fe-a438-fbf2ceececae] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 847.711682] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-954d9845-5977-47c6-bfa2-095783c80ce2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.717822] env[63531]: DEBUG oslo_vmware.api [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Waiting for the task: (returnval){ [ 847.717822] env[63531]: value = "task-1117938" [ 847.717822] env[63531]: _type = "Task" [ 847.717822] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.727043] env[63531]: DEBUG oslo_vmware.api [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': task-1117938, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.764791] env[63531]: DEBUG oslo_vmware.api [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': task-1117937, 'name': Rename_Task, 'duration_secs': 0.1827} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.765662] env[63531]: DEBUG nova.network.neutron [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Successfully created port: b9ea99ca-f03c-4164-b8ae-b68245e39a5a {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 847.767351] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 847.767492] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c1faba71-9551-42f0-bb98-55038e399d9b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.771203] env[63531]: DEBUG nova.compute.manager [req-4c489f9a-d58c-4108-99f8-7b3f8cc83eaf req-0bbf0019-3569-4cf1-89ee-eb562603749f service nova] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] Received event network-changed-e93a4c61-805f-4c14-8649-6718ca8f490d {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 847.771396] env[63531]: DEBUG nova.compute.manager [req-4c489f9a-d58c-4108-99f8-7b3f8cc83eaf req-0bbf0019-3569-4cf1-89ee-eb562603749f service nova] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] Refreshing instance network info cache due to event network-changed-e93a4c61-805f-4c14-8649-6718ca8f490d. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 847.771612] env[63531]: DEBUG oslo_concurrency.lockutils [req-4c489f9a-d58c-4108-99f8-7b3f8cc83eaf req-0bbf0019-3569-4cf1-89ee-eb562603749f service nova] Acquiring lock "refresh_cache-45b05c54-6865-47a1-ba93-90ad3e1ba07e" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.771757] env[63531]: DEBUG oslo_concurrency.lockutils [req-4c489f9a-d58c-4108-99f8-7b3f8cc83eaf req-0bbf0019-3569-4cf1-89ee-eb562603749f service nova] Acquired lock "refresh_cache-45b05c54-6865-47a1-ba93-90ad3e1ba07e" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.771921] env[63531]: DEBUG nova.network.neutron [req-4c489f9a-d58c-4108-99f8-7b3f8cc83eaf req-0bbf0019-3569-4cf1-89ee-eb562603749f service nova] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] Refreshing network info cache for port e93a4c61-805f-4c14-8649-6718ca8f490d {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 847.778651] env[63531]: DEBUG oslo_vmware.api [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Waiting for the task: (returnval){ [ 847.778651] env[63531]: value = "task-1117939" [ 847.778651] env[63531]: _type = "Task" [ 847.778651] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.789887] env[63531]: DEBUG oslo_vmware.api [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': task-1117939, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.853133] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquiring lock "37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.853456] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.869689] env[63531]: DEBUG oslo_vmware.api [None req-92007411-3bac-410d-a9c6-c0b9a37679fa tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1117934, 'name': RemoveSnapshot_Task, 'duration_secs': 0.676396} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.869995] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-92007411-3bac-410d-a9c6-c0b9a37679fa tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Deleted Snapshot of the VM instance {{(pid=63531) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 847.870260] env[63531]: INFO nova.compute.manager [None req-92007411-3bac-410d-a9c6-c0b9a37679fa tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Took 15.50 seconds to snapshot the instance on the hypervisor. [ 847.894466] env[63531]: DEBUG nova.compute.manager [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 847.989195] env[63531]: DEBUG nova.network.neutron [-] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.034478] env[63531]: DEBUG nova.network.neutron [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 848.115257] env[63531]: DEBUG nova.network.neutron [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Successfully created port: cbb5eb21-bc39-4608-820f-a77b3f5fdedf {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 848.137715] env[63531]: DEBUG oslo_vmware.api [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Task: {'id': task-1117935, 'name': PowerOnVM_Task, 'duration_secs': 0.573119} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.140159] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] [instance: 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 848.140386] env[63531]: INFO nova.compute.manager [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] [instance: 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d] Took 9.47 seconds to spawn the instance on the hypervisor. [ 848.140586] env[63531]: DEBUG nova.compute.manager [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] [instance: 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 848.141603] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9dbec51-bab9-4222-96a2-4e5856415b58 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.231103] env[63531]: DEBUG oslo_vmware.api [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': task-1117938, 'name': PowerOnVM_Task} progress is 87%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.290147] env[63531]: DEBUG oslo_vmware.api [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': task-1117939, 'name': PowerOnVM_Task} progress is 81%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.356157] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e341ac6-6359-4350-a318-4c5ccbf98d0d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.359070] env[63531]: DEBUG nova.compute.manager [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 848.364876] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf0de14a-f253-476b-89eb-7355c5b59cf8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.410030] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81e22f71-cb77-4a36-a32a-6b847a48c265 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.419407] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a59b94b5-6fe7-4080-aad8-785f51b3353c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.435932] env[63531]: DEBUG nova.compute.provider_tree [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 848.472143] env[63531]: DEBUG nova.compute.manager [None req-92007411-3bac-410d-a9c6-c0b9a37679fa tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Found 1 images (rotation: 2) {{(pid=63531) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 848.494409] env[63531]: INFO nova.compute.manager [-] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Took 1.60 seconds to deallocate network for instance. [ 848.596344] env[63531]: DEBUG nova.network.neutron [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Updating instance_info_cache with network_info: [{"id": "13365dfb-f81c-4e56-bf30-0a09dec4f561", "address": "fa:16:3e:2a:0b:33", "network": {"id": "bf99b8cb-9efb-4a74-9db1-0f5f0875f26a", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-998205595-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2e99349303e413b9f7896d449cb7dff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc6e6fe1-c4f5-4389-a49f-0978060eebb4", "external-id": "nsx-vlan-transportzone-829", "segmentation_id": 829, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap13365dfb-f8", "ovs_interfaceid": "13365dfb-f81c-4e56-bf30-0a09dec4f561", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.662479] env[63531]: INFO nova.compute.manager [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] [instance: 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d] Took 51.74 seconds to build instance. [ 848.728268] env[63531]: DEBUG oslo_vmware.api [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': task-1117938, 'name': PowerOnVM_Task, 'duration_secs': 0.92313} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.728611] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 3eeaa6f9-60d9-43fe-a438-fbf2ceececae] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 848.728877] env[63531]: INFO nova.compute.manager [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 3eeaa6f9-60d9-43fe-a438-fbf2ceececae] Took 7.45 seconds to spawn the instance on the hypervisor. [ 848.729085] env[63531]: DEBUG nova.compute.manager [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 3eeaa6f9-60d9-43fe-a438-fbf2ceececae] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 848.729880] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d4d2638-83ab-414b-9ea8-9fa94e795905 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.788561] env[63531]: DEBUG oslo_vmware.api [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': task-1117939, 'name': PowerOnVM_Task, 'duration_secs': 0.921243} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.788842] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 848.789101] env[63531]: INFO nova.compute.manager [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Took 5.07 seconds to spawn the instance on the hypervisor. [ 848.789338] env[63531]: DEBUG nova.compute.manager [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 848.790088] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b645a4f-b685-4c0c-b2e6-211458b4d82c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.887066] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.910319] env[63531]: DEBUG nova.compute.manager [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 848.936054] env[63531]: DEBUG nova.virt.hardware [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 848.936390] env[63531]: DEBUG nova.virt.hardware [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 848.936562] env[63531]: DEBUG nova.virt.hardware [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 848.936758] env[63531]: DEBUG nova.virt.hardware [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 848.936908] env[63531]: DEBUG nova.virt.hardware [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 848.937148] env[63531]: DEBUG nova.virt.hardware [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 848.937313] env[63531]: DEBUG nova.virt.hardware [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 848.937503] env[63531]: DEBUG nova.virt.hardware [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 848.937673] env[63531]: DEBUG nova.virt.hardware [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 848.937844] env[63531]: DEBUG nova.virt.hardware [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 848.938036] env[63531]: DEBUG nova.virt.hardware [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 848.938911] env[63531]: DEBUG nova.scheduler.client.report [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 848.943292] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53da278c-0abe-4aba-b2fe-002e34cd0da5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.947491] env[63531]: DEBUG nova.network.neutron [req-4c489f9a-d58c-4108-99f8-7b3f8cc83eaf req-0bbf0019-3569-4cf1-89ee-eb562603749f service nova] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] Updated VIF entry in instance network info cache for port e93a4c61-805f-4c14-8649-6718ca8f490d. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 848.947821] env[63531]: DEBUG nova.network.neutron [req-4c489f9a-d58c-4108-99f8-7b3f8cc83eaf req-0bbf0019-3569-4cf1-89ee-eb562603749f service nova] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] Updating instance_info_cache with network_info: [{"id": "e93a4c61-805f-4c14-8649-6718ca8f490d", "address": "fa:16:3e:63:6d:f8", "network": {"id": "4dfee301-fd80-46f1-8ed0-0499137e0bf1", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-780216327-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.174", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "95a8149e13534b3faa6c70be2302bfce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "29945b49-d2b8-449d-9531-437917f49839", "external-id": "nsx-vlan-transportzone-787", "segmentation_id": 787, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape93a4c61-80", "ovs_interfaceid": "e93a4c61-805f-4c14-8649-6718ca8f490d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.954335] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4238c7cc-3a01-4f64-8624-792e32bcbf5a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.002381] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a2338418-703e-4117-a6e8-2c9a448381e7 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.085253] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ae82578e-ba32-4675-8601-c68db67ff728 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "29dd6de3-2765-408c-acc0-da47e5e0a977" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.085569] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ae82578e-ba32-4675-8601-c68db67ff728 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "29dd6de3-2765-408c-acc0-da47e5e0a977" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.085833] env[63531]: INFO nova.compute.manager [None req-ae82578e-ba32-4675-8601-c68db67ff728 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Rebooting instance [ 849.098896] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Releasing lock "refresh_cache-2b0e4285-310a-439c-bc50-000a7c5ef7f9" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 849.098994] env[63531]: DEBUG nova.compute.manager [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Instance network_info: |[{"id": "13365dfb-f81c-4e56-bf30-0a09dec4f561", "address": "fa:16:3e:2a:0b:33", "network": {"id": "bf99b8cb-9efb-4a74-9db1-0f5f0875f26a", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-998205595-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2e99349303e413b9f7896d449cb7dff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc6e6fe1-c4f5-4389-a49f-0978060eebb4", "external-id": "nsx-vlan-transportzone-829", "segmentation_id": 829, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap13365dfb-f8", "ovs_interfaceid": "13365dfb-f81c-4e56-bf30-0a09dec4f561", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 849.099698] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2a:0b:33', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bc6e6fe1-c4f5-4389-a49f-0978060eebb4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '13365dfb-f81c-4e56-bf30-0a09dec4f561', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 849.107542] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Creating folder: Project (c2e99349303e413b9f7896d449cb7dff). Parent ref: group-v244585. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 849.110258] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ee721d1d-83c3-4fbb-987a-370ea2d8af8e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.119789] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Created folder: Project (c2e99349303e413b9f7896d449cb7dff) in parent group-v244585. [ 849.120036] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Creating folder: Instances. Parent ref: group-v244707. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 849.120317] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-716f88ec-e167-4d27-a6e4-fa5bbfc0abb2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.132047] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Created folder: Instances in parent group-v244707. [ 849.132327] env[63531]: DEBUG oslo.service.loopingcall [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 849.132541] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 849.132757] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dc3ace6d-7a5d-4613-be22-3d2103f73d35 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.152228] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 849.152228] env[63531]: value = "task-1117942" [ 849.152228] env[63531]: _type = "Task" [ 849.152228] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.161722] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117942, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.168375] env[63531]: DEBUG oslo_concurrency.lockutils [None req-71c10e17-944f-4a72-a4b6-7837f0c02757 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Lock "56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 53.251s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.245682] env[63531]: INFO nova.compute.manager [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 3eeaa6f9-60d9-43fe-a438-fbf2ceececae] Took 48.30 seconds to build instance. [ 849.309607] env[63531]: INFO nova.compute.manager [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Took 38.55 seconds to build instance. [ 849.447358] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.569s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.448049] env[63531]: DEBUG nova.compute.manager [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: db83adf4-6183-4330-b260-77d1f5daf899] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 849.450796] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ea661ef7-af43-46e1-812d-b8cc1057ecf2 tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.972s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.451330] env[63531]: DEBUG nova.objects.instance [None req-ea661ef7-af43-46e1-812d-b8cc1057ecf2 tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Lazy-loading 'resources' on Instance uuid eb6c84d7-e0e6-4999-9018-889b99dc67ad {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 849.452733] env[63531]: DEBUG oslo_concurrency.lockutils [req-4c489f9a-d58c-4108-99f8-7b3f8cc83eaf req-0bbf0019-3569-4cf1-89ee-eb562603749f service nova] Releasing lock "refresh_cache-45b05c54-6865-47a1-ba93-90ad3e1ba07e" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 849.605363] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ae82578e-ba32-4675-8601-c68db67ff728 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "refresh_cache-29dd6de3-2765-408c-acc0-da47e5e0a977" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.610826] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ae82578e-ba32-4675-8601-c68db67ff728 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquired lock "refresh_cache-29dd6de3-2765-408c-acc0-da47e5e0a977" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.610826] env[63531]: DEBUG nova.network.neutron [None req-ae82578e-ba32-4675-8601-c68db67ff728 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 849.662620] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117942, 'name': CreateVM_Task, 'duration_secs': 0.452907} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.663043] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 849.663788] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.663970] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.664328] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 849.665132] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e271ab06-8ce6-4598-9888-fc94e004178a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.670434] env[63531]: DEBUG oslo_vmware.api [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 849.670434] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]527019eb-b692-ed36-7d90-57d016155138" [ 849.670434] env[63531]: _type = "Task" [ 849.670434] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.678983] env[63531]: DEBUG oslo_vmware.api [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]527019eb-b692-ed36-7d90-57d016155138, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.735643] env[63531]: DEBUG nova.compute.manager [req-4d79020c-78c3-48b6-9357-8a27e35b569a req-da215877-58ff-45b3-a09b-046062dae847 service nova] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Received event network-changed-13365dfb-f81c-4e56-bf30-0a09dec4f561 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 849.735842] env[63531]: DEBUG nova.compute.manager [req-4d79020c-78c3-48b6-9357-8a27e35b569a req-da215877-58ff-45b3-a09b-046062dae847 service nova] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Refreshing instance network info cache due to event network-changed-13365dfb-f81c-4e56-bf30-0a09dec4f561. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 849.736084] env[63531]: DEBUG oslo_concurrency.lockutils [req-4d79020c-78c3-48b6-9357-8a27e35b569a req-da215877-58ff-45b3-a09b-046062dae847 service nova] Acquiring lock "refresh_cache-2b0e4285-310a-439c-bc50-000a7c5ef7f9" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.736915] env[63531]: DEBUG oslo_concurrency.lockutils [req-4d79020c-78c3-48b6-9357-8a27e35b569a req-da215877-58ff-45b3-a09b-046062dae847 service nova] Acquired lock "refresh_cache-2b0e4285-310a-439c-bc50-000a7c5ef7f9" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.736915] env[63531]: DEBUG nova.network.neutron [req-4d79020c-78c3-48b6-9357-8a27e35b569a req-da215877-58ff-45b3-a09b-046062dae847 service nova] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Refreshing network info cache for port 13365dfb-f81c-4e56-bf30-0a09dec4f561 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 849.749243] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ce79e6d7-e03d-4597-be0c-b36c1e6453fc tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Lock "3eeaa6f9-60d9-43fe-a438-fbf2ceececae" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 49.808s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.815217] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f03addee-0e84-47f1-a2ae-3f7efcd280b5 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Lock "09feb12a-da9f-4bb7-959d-84c5c7c58e99" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 49.477s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.831934] env[63531]: DEBUG nova.compute.manager [req-5a02486a-ad75-41d9-834b-ac38cccba6c2 req-4b510396-a84e-4a74-b1ed-985fca9b2f7f service nova] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Received event network-vif-deleted-96e03393-975e-4587-b22c-59fe3ba65718 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 849.851761] env[63531]: DEBUG nova.network.neutron [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Successfully updated port: b9ea99ca-f03c-4164-b8ae-b68245e39a5a {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 849.955171] env[63531]: DEBUG nova.compute.utils [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 849.957461] env[63531]: DEBUG nova.compute.manager [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: db83adf4-6183-4330-b260-77d1f5daf899] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 849.957658] env[63531]: DEBUG nova.network.neutron [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: db83adf4-6183-4330-b260-77d1f5daf899] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 850.015199] env[63531]: DEBUG nova.policy [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '30c2b29064b44c48ad7e76d445c65317', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '04a4c814245346a9b73253a25a822fef', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 850.061170] env[63531]: DEBUG nova.compute.manager [None req-dba9d971-47b5-433d-b369-13c4616060ba tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 850.062771] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69c97632-e39e-47a6-8a09-2ea66da09e8d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.188476] env[63531]: DEBUG oslo_vmware.api [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]527019eb-b692-ed36-7d90-57d016155138, 'name': SearchDatastore_Task, 'duration_secs': 0.022198} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.188842] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.189101] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 850.189349] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.189493] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.189674] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 850.189966] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5334efb1-a18f-43a0-af91-a4a13a724d30 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.200719] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 850.200935] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 850.204756] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e9684e46-9917-4a5f-ac1d-21b53ac9a817 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.211590] env[63531]: DEBUG oslo_vmware.api [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 850.211590] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5202ec23-225f-f026-2664-790cbf751725" [ 850.211590] env[63531]: _type = "Task" [ 850.211590] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.217344] env[63531]: INFO nova.compute.manager [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Rebuilding instance [ 850.230916] env[63531]: DEBUG oslo_vmware.api [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5202ec23-225f-f026-2664-790cbf751725, 'name': SearchDatastore_Task} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.230916] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3e3cc4f2-e5b2-4ea8-8834-1b82fc48ef14 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.234061] env[63531]: DEBUG oslo_vmware.api [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 850.234061] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]527b1bea-1483-0434-d349-4678af64bc0e" [ 850.234061] env[63531]: _type = "Task" [ 850.234061] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.252528] env[63531]: DEBUG oslo_vmware.api [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]527b1bea-1483-0434-d349-4678af64bc0e, 'name': SearchDatastore_Task, 'duration_secs': 0.00978} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.257636] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.257636] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] 2b0e4285-310a-439c-bc50-000a7c5ef7f9/2b0e4285-310a-439c-bc50-000a7c5ef7f9.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 850.257935] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1dd3e7ed-a866-4319-8f15-93be5dcf390e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.265427] env[63531]: DEBUG oslo_vmware.api [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 850.265427] env[63531]: value = "task-1117943" [ 850.265427] env[63531]: _type = "Task" [ 850.265427] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.284300] env[63531]: DEBUG oslo_vmware.api [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1117943, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.304891] env[63531]: DEBUG nova.compute.manager [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 850.306287] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27e50dd4-d9a0-48fd-ad70-d44ede41218e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.449894] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12a8e800-e27a-4900-a23a-2065203fa36e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.458853] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-864b10d1-8676-48ce-a0bb-0326de74e633 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.462842] env[63531]: DEBUG nova.compute.manager [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: db83adf4-6183-4330-b260-77d1f5daf899] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 850.511747] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f254f95-c0cd-4783-bcaf-a75ec5faa1b1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.519067] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf975a41-66a9-4f86-acd0-ef99447a14fd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.538067] env[63531]: DEBUG nova.compute.provider_tree [None req-ea661ef7-af43-46e1-812d-b8cc1057ecf2 tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 850.578527] env[63531]: INFO nova.compute.manager [None req-dba9d971-47b5-433d-b369-13c4616060ba tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] instance snapshotting [ 850.579383] env[63531]: DEBUG nova.objects.instance [None req-dba9d971-47b5-433d-b369-13c4616060ba tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lazy-loading 'flavor' on Instance uuid 9a3fad46-a15a-451c-bdab-a3c8cc8add07 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 850.744289] env[63531]: DEBUG nova.network.neutron [None req-ae82578e-ba32-4675-8601-c68db67ff728 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Updating instance_info_cache with network_info: [{"id": "262abfb7-7bcf-40e6-9ce3-b502de45b3b4", "address": "fa:16:3e:0e:00:c9", "network": {"id": "0667c176-cfdf-4e43-b152-e00bdc224448", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-243715985-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.217", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "308eabafea9c4188a58a0f1c22074d2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8f580e6-1d86-41ee-9ebe-c531cb9299c6", "external-id": "nsx-vlan-transportzone-150", "segmentation_id": 150, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap262abfb7-7b", "ovs_interfaceid": "262abfb7-7bcf-40e6-9ce3-b502de45b3b4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.761375] env[63531]: DEBUG nova.network.neutron [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: db83adf4-6183-4330-b260-77d1f5daf899] Successfully created port: 895f47c1-21ad-45cc-b1d4-e7a8fa7398cb {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 850.779635] env[63531]: DEBUG oslo_vmware.api [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1117943, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.823207] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 850.823207] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5f704315-2918-463b-89e0-03a2a917e156 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.832875] env[63531]: DEBUG oslo_vmware.api [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Waiting for the task: (returnval){ [ 850.832875] env[63531]: value = "task-1117944" [ 850.832875] env[63531]: _type = "Task" [ 850.832875] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.833807] env[63531]: DEBUG nova.network.neutron [req-4d79020c-78c3-48b6-9357-8a27e35b569a req-da215877-58ff-45b3-a09b-046062dae847 service nova] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Updated VIF entry in instance network info cache for port 13365dfb-f81c-4e56-bf30-0a09dec4f561. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 850.834392] env[63531]: DEBUG nova.network.neutron [req-4d79020c-78c3-48b6-9357-8a27e35b569a req-da215877-58ff-45b3-a09b-046062dae847 service nova] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Updating instance_info_cache with network_info: [{"id": "13365dfb-f81c-4e56-bf30-0a09dec4f561", "address": "fa:16:3e:2a:0b:33", "network": {"id": "bf99b8cb-9efb-4a74-9db1-0f5f0875f26a", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-998205595-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2e99349303e413b9f7896d449cb7dff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc6e6fe1-c4f5-4389-a49f-0978060eebb4", "external-id": "nsx-vlan-transportzone-829", "segmentation_id": 829, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap13365dfb-f8", "ovs_interfaceid": "13365dfb-f81c-4e56-bf30-0a09dec4f561", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.852798] env[63531]: DEBUG oslo_vmware.api [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': task-1117944, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.044020] env[63531]: DEBUG nova.scheduler.client.report [None req-ea661ef7-af43-46e1-812d-b8cc1057ecf2 tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 851.085899] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f932fc35-5e63-43c0-a57f-bf6a5df0b8df {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.112351] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37091563-2563-4554-aee8-b72551b816b6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.248269] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ae82578e-ba32-4675-8601-c68db67ff728 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Releasing lock "refresh_cache-29dd6de3-2765-408c-acc0-da47e5e0a977" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.251350] env[63531]: DEBUG nova.compute.manager [None req-ae82578e-ba32-4675-8601-c68db67ff728 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 851.252684] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5693f79-ded1-4e74-a9e4-e3042e92c017 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.280779] env[63531]: DEBUG oslo_vmware.api [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1117943, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.616569} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.281177] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] 2b0e4285-310a-439c-bc50-000a7c5ef7f9/2b0e4285-310a-439c-bc50-000a7c5ef7f9.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 851.281486] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 851.281753] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6d2c8d9a-5a9c-4029-9d97-5399ec83ce8f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.289028] env[63531]: DEBUG oslo_vmware.api [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 851.289028] env[63531]: value = "task-1117945" [ 851.289028] env[63531]: _type = "Task" [ 851.289028] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.297342] env[63531]: DEBUG oslo_vmware.api [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1117945, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.348846] env[63531]: DEBUG oslo_concurrency.lockutils [req-4d79020c-78c3-48b6-9357-8a27e35b569a req-da215877-58ff-45b3-a09b-046062dae847 service nova] Releasing lock "refresh_cache-2b0e4285-310a-439c-bc50-000a7c5ef7f9" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.349333] env[63531]: DEBUG oslo_vmware.api [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': task-1117944, 'name': PowerOffVM_Task, 'duration_secs': 0.272604} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.349932] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 851.349932] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 851.350686] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45b10d2d-b925-4e27-b24d-a5caca8caa81 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.357535] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 851.357832] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-caa8ee12-7f20-4dea-b6b7-004bd49480cf {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.383700] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 851.383982] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 851.384435] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Deleting the datastore file [datastore1] 09feb12a-da9f-4bb7-959d-84c5c7c58e99 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 851.385505] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8d700476-ff45-4036-aa52-556dc3b06442 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.392808] env[63531]: DEBUG oslo_vmware.api [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Waiting for the task: (returnval){ [ 851.392808] env[63531]: value = "task-1117947" [ 851.392808] env[63531]: _type = "Task" [ 851.392808] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.401975] env[63531]: DEBUG oslo_vmware.api [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': task-1117947, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.477522] env[63531]: DEBUG nova.compute.manager [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: db83adf4-6183-4330-b260-77d1f5daf899] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 851.507589] env[63531]: DEBUG nova.virt.hardware [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='80c8f631c5725a4f3d3055c5d3c53dfd',container_format='bare',created_at=2024-10-03T08:20:29Z,direct_url=,disk_format='vmdk',id=8197d52e-21f1-413d-8ef3-59307facd73d,min_disk=1,min_ram=0,name='tempest-test-snap-22594613',owner='04a4c814245346a9b73253a25a822fef',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2024-10-03T08:20:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 851.507856] env[63531]: DEBUG nova.virt.hardware [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 851.508047] env[63531]: DEBUG nova.virt.hardware [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 851.508271] env[63531]: DEBUG nova.virt.hardware [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 851.508471] env[63531]: DEBUG nova.virt.hardware [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 851.508640] env[63531]: DEBUG nova.virt.hardware [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 851.508874] env[63531]: DEBUG nova.virt.hardware [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 851.509056] env[63531]: DEBUG nova.virt.hardware [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 851.509241] env[63531]: DEBUG nova.virt.hardware [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 851.509438] env[63531]: DEBUG nova.virt.hardware [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 851.509684] env[63531]: DEBUG nova.virt.hardware [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 851.510677] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a556551-ccff-4171-ba70-2193784b267a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.519283] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ab8547c-6cc2-4d07-9970-f4ab326c352f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.548759] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ea661ef7-af43-46e1-812d-b8cc1057ecf2 tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.098s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.551927] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.683s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.553974] env[63531]: INFO nova.compute.claims [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 851.579917] env[63531]: INFO nova.scheduler.client.report [None req-ea661ef7-af43-46e1-812d-b8cc1057ecf2 tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Deleted allocations for instance eb6c84d7-e0e6-4999-9018-889b99dc67ad [ 851.624619] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-dba9d971-47b5-433d-b369-13c4616060ba tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Creating Snapshot of the VM instance {{(pid=63531) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 851.625319] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-80ce2d3d-5077-4889-94e4-1029a64ba025 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.633308] env[63531]: DEBUG oslo_vmware.api [None req-dba9d971-47b5-433d-b369-13c4616060ba tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 851.633308] env[63531]: value = "task-1117948" [ 851.633308] env[63531]: _type = "Task" [ 851.633308] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.647555] env[63531]: DEBUG oslo_vmware.api [None req-dba9d971-47b5-433d-b369-13c4616060ba tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1117948, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.798575] env[63531]: DEBUG oslo_vmware.api [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1117945, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.215709} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.799951] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 851.801087] env[63531]: DEBUG nova.compute.manager [req-329a54d2-ca20-4997-8156-f4067dcd3a69 req-8595ac04-cfe4-4752-ad9d-04ede14d43aa service nova] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Received event network-vif-plugged-b9ea99ca-f03c-4164-b8ae-b68245e39a5a {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 851.801302] env[63531]: DEBUG oslo_concurrency.lockutils [req-329a54d2-ca20-4997-8156-f4067dcd3a69 req-8595ac04-cfe4-4752-ad9d-04ede14d43aa service nova] Acquiring lock "92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.801537] env[63531]: DEBUG oslo_concurrency.lockutils [req-329a54d2-ca20-4997-8156-f4067dcd3a69 req-8595ac04-cfe4-4752-ad9d-04ede14d43aa service nova] Lock "92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.801710] env[63531]: DEBUG oslo_concurrency.lockutils [req-329a54d2-ca20-4997-8156-f4067dcd3a69 req-8595ac04-cfe4-4752-ad9d-04ede14d43aa service nova] Lock "92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.801877] env[63531]: DEBUG nova.compute.manager [req-329a54d2-ca20-4997-8156-f4067dcd3a69 req-8595ac04-cfe4-4752-ad9d-04ede14d43aa service nova] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] No waiting events found dispatching network-vif-plugged-b9ea99ca-f03c-4164-b8ae-b68245e39a5a {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 851.802057] env[63531]: WARNING nova.compute.manager [req-329a54d2-ca20-4997-8156-f4067dcd3a69 req-8595ac04-cfe4-4752-ad9d-04ede14d43aa service nova] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Received unexpected event network-vif-plugged-b9ea99ca-f03c-4164-b8ae-b68245e39a5a for instance with vm_state building and task_state spawning. [ 851.802295] env[63531]: DEBUG nova.compute.manager [req-329a54d2-ca20-4997-8156-f4067dcd3a69 req-8595ac04-cfe4-4752-ad9d-04ede14d43aa service nova] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Received event network-changed-b9ea99ca-f03c-4164-b8ae-b68245e39a5a {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 851.802459] env[63531]: DEBUG nova.compute.manager [req-329a54d2-ca20-4997-8156-f4067dcd3a69 req-8595ac04-cfe4-4752-ad9d-04ede14d43aa service nova] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Refreshing instance network info cache due to event network-changed-b9ea99ca-f03c-4164-b8ae-b68245e39a5a. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 851.802890] env[63531]: DEBUG oslo_concurrency.lockutils [req-329a54d2-ca20-4997-8156-f4067dcd3a69 req-8595ac04-cfe4-4752-ad9d-04ede14d43aa service nova] Acquiring lock "refresh_cache-92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.802890] env[63531]: DEBUG oslo_concurrency.lockutils [req-329a54d2-ca20-4997-8156-f4067dcd3a69 req-8595ac04-cfe4-4752-ad9d-04ede14d43aa service nova] Acquired lock "refresh_cache-92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.802995] env[63531]: DEBUG nova.network.neutron [req-329a54d2-ca20-4997-8156-f4067dcd3a69 req-8595ac04-cfe4-4752-ad9d-04ede14d43aa service nova] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Refreshing network info cache for port b9ea99ca-f03c-4164-b8ae-b68245e39a5a {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 851.804470] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1aed585-79f3-4d12-8340-b4bb4ab8f1c8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.835558] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Reconfiguring VM instance instance-00000043 to attach disk [datastore2] 2b0e4285-310a-439c-bc50-000a7c5ef7f9/2b0e4285-310a-439c-bc50-000a7c5ef7f9.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 851.836481] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f80d79b9-6e7b-4b6e-a9c4-ca768d32eaf7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.862986] env[63531]: DEBUG oslo_vmware.api [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 851.862986] env[63531]: value = "task-1117949" [ 851.862986] env[63531]: _type = "Task" [ 851.862986] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.871691] env[63531]: DEBUG oslo_vmware.api [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1117949, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.902640] env[63531]: DEBUG oslo_vmware.api [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': task-1117947, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.111187} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.902887] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 851.903093] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 851.903279] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 852.091743] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ea661ef7-af43-46e1-812d-b8cc1057ecf2 tempest-InstanceActionsNegativeTestJSON-1912499313 tempest-InstanceActionsNegativeTestJSON-1912499313-project-member] Lock "eb6c84d7-e0e6-4999-9018-889b99dc67ad" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.621s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.145610] env[63531]: DEBUG oslo_vmware.api [None req-dba9d971-47b5-433d-b369-13c4616060ba tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1117948, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.285273] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b85e384-d7ab-4a7f-a95f-41b600bf09c7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.293264] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-ae82578e-ba32-4675-8601-c68db67ff728 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Doing hard reboot of VM {{(pid=63531) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1064}} [ 852.293264] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-c239fa96-45c7-46bb-9eb0-3f651136ae65 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.307378] env[63531]: DEBUG oslo_vmware.api [None req-ae82578e-ba32-4675-8601-c68db67ff728 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 852.307378] env[63531]: value = "task-1117950" [ 852.307378] env[63531]: _type = "Task" [ 852.307378] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.321246] env[63531]: DEBUG oslo_vmware.api [None req-ae82578e-ba32-4675-8601-c68db67ff728 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1117950, 'name': ResetVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.363151] env[63531]: DEBUG nova.network.neutron [req-329a54d2-ca20-4997-8156-f4067dcd3a69 req-8595ac04-cfe4-4752-ad9d-04ede14d43aa service nova] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 852.377025] env[63531]: DEBUG oslo_vmware.api [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1117949, 'name': ReconfigVM_Task, 'duration_secs': 0.357752} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.377025] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1654cc47-4170-4ce7-ab39-6549bb1aa779 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Acquiring lock "56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.377025] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1654cc47-4170-4ce7-ab39-6549bb1aa779 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Lock "56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.377025] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1654cc47-4170-4ce7-ab39-6549bb1aa779 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Acquiring lock "56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.377025] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1654cc47-4170-4ce7-ab39-6549bb1aa779 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Lock "56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.377025] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1654cc47-4170-4ce7-ab39-6549bb1aa779 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Lock "56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.379289] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Reconfigured VM instance instance-00000043 to attach disk [datastore2] 2b0e4285-310a-439c-bc50-000a7c5ef7f9/2b0e4285-310a-439c-bc50-000a7c5ef7f9.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 852.380445] env[63531]: INFO nova.compute.manager [None req-1654cc47-4170-4ce7-ab39-6549bb1aa779 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] [instance: 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d] Terminating instance [ 852.383956] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b9f86c47-247c-4497-b18d-123f35523fa7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.384913] env[63531]: DEBUG nova.compute.manager [None req-1654cc47-4170-4ce7-ab39-6549bb1aa779 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] [instance: 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 852.385425] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-1654cc47-4170-4ce7-ab39-6549bb1aa779 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] [instance: 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 852.386328] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33d863b6-9038-4196-965a-48bb8cfb2abe {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.395254] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-1654cc47-4170-4ce7-ab39-6549bb1aa779 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] [instance: 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 852.396470] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-13ccbbe0-8f51-4c7f-9b2d-5dde82e53872 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.398286] env[63531]: DEBUG oslo_vmware.api [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 852.398286] env[63531]: value = "task-1117951" [ 852.398286] env[63531]: _type = "Task" [ 852.398286] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.402931] env[63531]: DEBUG oslo_vmware.api [None req-1654cc47-4170-4ce7-ab39-6549bb1aa779 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Waiting for the task: (returnval){ [ 852.402931] env[63531]: value = "task-1117952" [ 852.402931] env[63531]: _type = "Task" [ 852.402931] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.413601] env[63531]: DEBUG oslo_vmware.api [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1117951, 'name': Rename_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.419716] env[63531]: DEBUG oslo_vmware.api [None req-1654cc47-4170-4ce7-ab39-6549bb1aa779 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Task: {'id': task-1117952, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.508976] env[63531]: DEBUG nova.network.neutron [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Successfully updated port: cbb5eb21-bc39-4608-820f-a77b3f5fdedf {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 852.617341] env[63531]: DEBUG nova.network.neutron [req-329a54d2-ca20-4997-8156-f4067dcd3a69 req-8595ac04-cfe4-4752-ad9d-04ede14d43aa service nova] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.648750] env[63531]: DEBUG oslo_vmware.api [None req-dba9d971-47b5-433d-b369-13c4616060ba tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1117948, 'name': CreateSnapshot_Task, 'duration_secs': 0.881855} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.649817] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-dba9d971-47b5-433d-b369-13c4616060ba tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Created Snapshot of the VM instance {{(pid=63531) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 852.651157] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50352db5-00e5-49f7-9d66-151502ff1c94 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.746768] env[63531]: DEBUG nova.compute.manager [req-1edbfa8d-e976-4a94-905f-48d3b9b9c666 req-2c979459-e006-4a15-b114-de84454990b5 service nova] [instance: db83adf4-6183-4330-b260-77d1f5daf899] Received event network-vif-plugged-895f47c1-21ad-45cc-b1d4-e7a8fa7398cb {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 852.747242] env[63531]: DEBUG oslo_concurrency.lockutils [req-1edbfa8d-e976-4a94-905f-48d3b9b9c666 req-2c979459-e006-4a15-b114-de84454990b5 service nova] Acquiring lock "db83adf4-6183-4330-b260-77d1f5daf899-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.747710] env[63531]: DEBUG oslo_concurrency.lockutils [req-1edbfa8d-e976-4a94-905f-48d3b9b9c666 req-2c979459-e006-4a15-b114-de84454990b5 service nova] Lock "db83adf4-6183-4330-b260-77d1f5daf899-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.747907] env[63531]: DEBUG oslo_concurrency.lockutils [req-1edbfa8d-e976-4a94-905f-48d3b9b9c666 req-2c979459-e006-4a15-b114-de84454990b5 service nova] Lock "db83adf4-6183-4330-b260-77d1f5daf899-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.748115] env[63531]: DEBUG nova.compute.manager [req-1edbfa8d-e976-4a94-905f-48d3b9b9c666 req-2c979459-e006-4a15-b114-de84454990b5 service nova] [instance: db83adf4-6183-4330-b260-77d1f5daf899] No waiting events found dispatching network-vif-plugged-895f47c1-21ad-45cc-b1d4-e7a8fa7398cb {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 852.748573] env[63531]: WARNING nova.compute.manager [req-1edbfa8d-e976-4a94-905f-48d3b9b9c666 req-2c979459-e006-4a15-b114-de84454990b5 service nova] [instance: db83adf4-6183-4330-b260-77d1f5daf899] Received unexpected event network-vif-plugged-895f47c1-21ad-45cc-b1d4-e7a8fa7398cb for instance with vm_state building and task_state spawning. [ 852.825726] env[63531]: DEBUG oslo_vmware.api [None req-ae82578e-ba32-4675-8601-c68db67ff728 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1117950, 'name': ResetVM_Task, 'duration_secs': 0.14239} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.826044] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-ae82578e-ba32-4675-8601-c68db67ff728 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Did hard reboot of VM {{(pid=63531) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1068}} [ 852.826245] env[63531]: DEBUG nova.compute.manager [None req-ae82578e-ba32-4675-8601-c68db67ff728 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 852.827139] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b9444c8-f8ff-481e-b81c-fe6172a41bce {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.919846] env[63531]: DEBUG oslo_vmware.api [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1117951, 'name': Rename_Task, 'duration_secs': 0.144104} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.926017] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 852.926017] env[63531]: DEBUG oslo_vmware.api [None req-1654cc47-4170-4ce7-ab39-6549bb1aa779 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Task: {'id': task-1117952, 'name': PowerOffVM_Task, 'duration_secs': 0.193905} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.926017] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f6cc5932-b920-4f7c-9bb8-4b794cc26ebb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.927578] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-1654cc47-4170-4ce7-ab39-6549bb1aa779 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] [instance: 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 852.927910] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-1654cc47-4170-4ce7-ab39-6549bb1aa779 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] [instance: 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 852.927977] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3c8b3eaf-3ee7-4f08-83a5-99497683dcb8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.935712] env[63531]: DEBUG oslo_vmware.api [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 852.935712] env[63531]: value = "task-1117953" [ 852.935712] env[63531]: _type = "Task" [ 852.935712] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.952550] env[63531]: DEBUG oslo_vmware.api [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1117953, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.955916] env[63531]: DEBUG nova.virt.hardware [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 852.956239] env[63531]: DEBUG nova.virt.hardware [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 852.956457] env[63531]: DEBUG nova.virt.hardware [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 852.956585] env[63531]: DEBUG nova.virt.hardware [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 852.956734] env[63531]: DEBUG nova.virt.hardware [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 852.956886] env[63531]: DEBUG nova.virt.hardware [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 852.957340] env[63531]: DEBUG nova.virt.hardware [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 852.957543] env[63531]: DEBUG nova.virt.hardware [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 852.957726] env[63531]: DEBUG nova.virt.hardware [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 852.957894] env[63531]: DEBUG nova.virt.hardware [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 852.958738] env[63531]: DEBUG nova.virt.hardware [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 852.959945] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7577a4da-6162-4b12-b18b-abfc4cf6393f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.972450] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7042839b-4650-4564-af70-f29421e4a034 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.990972] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Instance VIF info [] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 852.997941] env[63531]: DEBUG oslo.service.loopingcall [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 853.001154] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 853.002829] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-48bb04d7-fff0-444f-a12f-a7dacad446e5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.017719] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-1654cc47-4170-4ce7-ab39-6549bb1aa779 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] [instance: 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 853.017719] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-1654cc47-4170-4ce7-ab39-6549bb1aa779 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] [instance: 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 853.017719] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-1654cc47-4170-4ce7-ab39-6549bb1aa779 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Deleting the datastore file [datastore1] 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 853.018046] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Acquiring lock "refresh_cache-92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.018275] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-26d661a2-f952-4cbe-9c12-112c45930f3e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.025494] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 853.025494] env[63531]: value = "task-1117955" [ 853.025494] env[63531]: _type = "Task" [ 853.025494] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.026897] env[63531]: DEBUG oslo_vmware.api [None req-1654cc47-4170-4ce7-ab39-6549bb1aa779 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Waiting for the task: (returnval){ [ 853.026897] env[63531]: value = "task-1117956" [ 853.026897] env[63531]: _type = "Task" [ 853.026897] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.037899] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117955, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.043547] env[63531]: DEBUG oslo_vmware.api [None req-1654cc47-4170-4ce7-ab39-6549bb1aa779 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Task: {'id': task-1117956, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.056071] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01a1a68d-2eb9-460c-85af-126b5a6e643b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.064764] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a15d3f09-3439-4f69-8953-2c044dfa9b5e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.095612] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0ee4df1-388d-447a-9ddf-9bb67eef918a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.103034] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f1f118a-895d-4b14-b462-c9f60ca319a9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.117061] env[63531]: DEBUG nova.compute.provider_tree [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 853.118639] env[63531]: DEBUG oslo_concurrency.lockutils [req-329a54d2-ca20-4997-8156-f4067dcd3a69 req-8595ac04-cfe4-4752-ad9d-04ede14d43aa service nova] Releasing lock "refresh_cache-92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.119208] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Acquired lock "refresh_cache-92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.119377] env[63531]: DEBUG nova.network.neutron [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 853.173896] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-dba9d971-47b5-433d-b369-13c4616060ba tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Creating linked-clone VM from snapshot {{(pid=63531) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 853.174284] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-8c1c85ac-f303-45d9-9f27-52826188578b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.184801] env[63531]: DEBUG oslo_vmware.api [None req-dba9d971-47b5-433d-b369-13c4616060ba tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 853.184801] env[63531]: value = "task-1117957" [ 853.184801] env[63531]: _type = "Task" [ 853.184801] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.193024] env[63531]: DEBUG oslo_vmware.api [None req-dba9d971-47b5-433d-b369-13c4616060ba tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1117957, 'name': CloneVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.232640] env[63531]: DEBUG nova.network.neutron [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: db83adf4-6183-4330-b260-77d1f5daf899] Successfully updated port: 895f47c1-21ad-45cc-b1d4-e7a8fa7398cb {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 853.307561] env[63531]: DEBUG oslo_concurrency.lockutils [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Acquiring lock "2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.307561] env[63531]: DEBUG oslo_concurrency.lockutils [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Lock "2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.307561] env[63531]: INFO nova.compute.manager [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Shelving [ 853.343957] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ae82578e-ba32-4675-8601-c68db67ff728 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "29dd6de3-2765-408c-acc0-da47e5e0a977" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.257s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.447791] env[63531]: DEBUG oslo_vmware.api [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1117953, 'name': PowerOnVM_Task, 'duration_secs': 0.493106} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.448134] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 853.448627] env[63531]: INFO nova.compute.manager [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Took 7.18 seconds to spawn the instance on the hypervisor. [ 853.448849] env[63531]: DEBUG nova.compute.manager [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 853.449972] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff9be943-b4f4-4833-a623-6f8c41713ab1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.542210] env[63531]: DEBUG oslo_vmware.api [None req-1654cc47-4170-4ce7-ab39-6549bb1aa779 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Task: {'id': task-1117956, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.213865} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.545873] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-1654cc47-4170-4ce7-ab39-6549bb1aa779 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 853.546314] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-1654cc47-4170-4ce7-ab39-6549bb1aa779 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] [instance: 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 853.546314] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-1654cc47-4170-4ce7-ab39-6549bb1aa779 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] [instance: 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 853.546535] env[63531]: INFO nova.compute.manager [None req-1654cc47-4170-4ce7-ab39-6549bb1aa779 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] [instance: 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d] Took 1.16 seconds to destroy the instance on the hypervisor. [ 853.546685] env[63531]: DEBUG oslo.service.loopingcall [None req-1654cc47-4170-4ce7-ab39-6549bb1aa779 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 853.547271] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117955, 'name': CreateVM_Task, 'duration_secs': 0.382501} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.547521] env[63531]: DEBUG nova.compute.manager [-] [instance: 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 853.547580] env[63531]: DEBUG nova.network.neutron [-] [instance: 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 853.550472] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 853.551646] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.552257] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.552330] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 853.552933] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a2f924d5-c744-4a99-8fa1-514f9788a3aa {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.557803] env[63531]: DEBUG oslo_vmware.api [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Waiting for the task: (returnval){ [ 853.557803] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5297da71-b866-af5c-8b0a-1598d23e072a" [ 853.557803] env[63531]: _type = "Task" [ 853.557803] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.566478] env[63531]: DEBUG oslo_vmware.api [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5297da71-b866-af5c-8b0a-1598d23e072a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.622582] env[63531]: DEBUG nova.scheduler.client.report [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 853.678912] env[63531]: DEBUG nova.network.neutron [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 853.701744] env[63531]: DEBUG oslo_vmware.api [None req-dba9d971-47b5-433d-b369-13c4616060ba tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1117957, 'name': CloneVM_Task} progress is 94%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.736068] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Acquiring lock "refresh_cache-db83adf4-6183-4330-b260-77d1f5daf899" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.736548] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Acquired lock "refresh_cache-db83adf4-6183-4330-b260-77d1f5daf899" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.736548] env[63531]: DEBUG nova.network.neutron [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: db83adf4-6183-4330-b260-77d1f5daf899] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 853.812745] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 853.813049] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ca352aa1-8e4f-4b65-946e-fc8e6f508ff2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.822591] env[63531]: DEBUG oslo_vmware.api [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Waiting for the task: (returnval){ [ 853.822591] env[63531]: value = "task-1117958" [ 853.822591] env[63531]: _type = "Task" [ 853.822591] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.830929] env[63531]: DEBUG oslo_vmware.api [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': task-1117958, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.979236] env[63531]: INFO nova.compute.manager [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Took 40.32 seconds to build instance. [ 854.068904] env[63531]: DEBUG oslo_vmware.api [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5297da71-b866-af5c-8b0a-1598d23e072a, 'name': SearchDatastore_Task, 'duration_secs': 0.010846} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.069323] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.069607] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 854.069980] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.070240] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.071719] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 854.071719] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ac164d9e-ee37-4318-942e-7cb535c4cffc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.082767] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 854.083195] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 854.083724] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7ad72a44-606c-4ab5-b589-0b13dd36e5f4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.090110] env[63531]: DEBUG oslo_vmware.api [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Waiting for the task: (returnval){ [ 854.090110] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52771009-2efa-fcba-e875-31ddd304c09f" [ 854.090110] env[63531]: _type = "Task" [ 854.090110] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.098725] env[63531]: DEBUG oslo_vmware.api [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52771009-2efa-fcba-e875-31ddd304c09f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.131139] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.579s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.131682] env[63531]: DEBUG nova.compute.manager [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 854.134180] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.976s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.137631] env[63531]: INFO nova.compute.claims [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 854.195914] env[63531]: DEBUG oslo_vmware.api [None req-dba9d971-47b5-433d-b369-13c4616060ba tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1117957, 'name': CloneVM_Task} progress is 94%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.289243] env[63531]: DEBUG nova.compute.manager [req-557089e7-fedb-4711-a820-e43883279ec4 req-f4231da1-4e4e-453a-a990-7eb07c8b97f9 service nova] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Received event network-vif-plugged-cbb5eb21-bc39-4608-820f-a77b3f5fdedf {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 854.289243] env[63531]: DEBUG oslo_concurrency.lockutils [req-557089e7-fedb-4711-a820-e43883279ec4 req-f4231da1-4e4e-453a-a990-7eb07c8b97f9 service nova] Acquiring lock "92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.289243] env[63531]: DEBUG oslo_concurrency.lockutils [req-557089e7-fedb-4711-a820-e43883279ec4 req-f4231da1-4e4e-453a-a990-7eb07c8b97f9 service nova] Lock "92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.289937] env[63531]: DEBUG oslo_concurrency.lockutils [req-557089e7-fedb-4711-a820-e43883279ec4 req-f4231da1-4e4e-453a-a990-7eb07c8b97f9 service nova] Lock "92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.290603] env[63531]: DEBUG nova.compute.manager [req-557089e7-fedb-4711-a820-e43883279ec4 req-f4231da1-4e4e-453a-a990-7eb07c8b97f9 service nova] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] No waiting events found dispatching network-vif-plugged-cbb5eb21-bc39-4608-820f-a77b3f5fdedf {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 854.291151] env[63531]: WARNING nova.compute.manager [req-557089e7-fedb-4711-a820-e43883279ec4 req-f4231da1-4e4e-453a-a990-7eb07c8b97f9 service nova] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Received unexpected event network-vif-plugged-cbb5eb21-bc39-4608-820f-a77b3f5fdedf for instance with vm_state building and task_state spawning. [ 854.291742] env[63531]: DEBUG nova.compute.manager [req-557089e7-fedb-4711-a820-e43883279ec4 req-f4231da1-4e4e-453a-a990-7eb07c8b97f9 service nova] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Received event network-changed-cbb5eb21-bc39-4608-820f-a77b3f5fdedf {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 854.292179] env[63531]: DEBUG nova.compute.manager [req-557089e7-fedb-4711-a820-e43883279ec4 req-f4231da1-4e4e-453a-a990-7eb07c8b97f9 service nova] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Refreshing instance network info cache due to event network-changed-cbb5eb21-bc39-4608-820f-a77b3f5fdedf. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 854.292678] env[63531]: DEBUG oslo_concurrency.lockutils [req-557089e7-fedb-4711-a820-e43883279ec4 req-f4231da1-4e4e-453a-a990-7eb07c8b97f9 service nova] Acquiring lock "refresh_cache-92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.319196] env[63531]: DEBUG nova.network.neutron [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: db83adf4-6183-4330-b260-77d1f5daf899] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 854.333125] env[63531]: DEBUG oslo_vmware.api [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': task-1117958, 'name': PowerOffVM_Task, 'duration_secs': 0.260996} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.333500] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 854.334198] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c192d398-bc8c-489d-96aa-5c3d71c362d4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.354610] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-763915ec-9915-4351-a27c-c661e134a8c3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.459657] env[63531]: DEBUG nova.network.neutron [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Updating instance_info_cache with network_info: [{"id": "b9ea99ca-f03c-4164-b8ae-b68245e39a5a", "address": "fa:16:3e:cf:a3:47", "network": {"id": "421a2d81-9456-47a7-996f-e15b955ef3d5", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-930999152", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.27", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bebde0e281d485ebdd70a1a3b613632", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7a0d5af-5be9-477a-837c-58ef55c717f4", "external-id": "nsx-vlan-transportzone-598", "segmentation_id": 598, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb9ea99ca-f0", "ovs_interfaceid": "b9ea99ca-f03c-4164-b8ae-b68245e39a5a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "cbb5eb21-bc39-4608-820f-a77b3f5fdedf", "address": "fa:16:3e:d0:eb:72", "network": {"id": "4ae7e465-bcda-457c-a060-64abd40b7152", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-251859226", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.59", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "8bebde0e281d485ebdd70a1a3b613632", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1dc3a0d3-9578-4c45-bc16-13ed5b84f5b4", "external-id": "nsx-vlan-transportzone-817", "segmentation_id": 817, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcbb5eb21-bc", "ovs_interfaceid": "cbb5eb21-bc39-4608-820f-a77b3f5fdedf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.481687] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c75e46f5-eb10-45ba-b38e-439f46117c7f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "2b0e4285-310a-439c-bc50-000a7c5ef7f9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 50.471s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.600655] env[63531]: DEBUG nova.network.neutron [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: db83adf4-6183-4330-b260-77d1f5daf899] Updating instance_info_cache with network_info: [{"id": "895f47c1-21ad-45cc-b1d4-e7a8fa7398cb", "address": "fa:16:3e:07:68:a6", "network": {"id": "887fff70-ea19-458d-a945-e4f3ed6dc185", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2099566288-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "04a4c814245346a9b73253a25a822fef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49a4d142-3f97-47fe-b074-58923c46815e", "external-id": "nsx-vlan-transportzone-565", "segmentation_id": 565, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap895f47c1-21", "ovs_interfaceid": "895f47c1-21ad-45cc-b1d4-e7a8fa7398cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.615490] env[63531]: DEBUG nova.network.neutron [-] [instance: 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.619746] env[63531]: DEBUG oslo_vmware.api [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52771009-2efa-fcba-e875-31ddd304c09f, 'name': SearchDatastore_Task, 'duration_secs': 0.009805} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.623275] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1acf0e65-4f74-4439-b00e-c3bdbd9b018a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.634614] env[63531]: DEBUG oslo_vmware.api [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Waiting for the task: (returnval){ [ 854.634614] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5253c480-9f71-71c6-20e7-9576b01d048e" [ 854.634614] env[63531]: _type = "Task" [ 854.634614] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.643602] env[63531]: DEBUG nova.compute.utils [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 854.652596] env[63531]: DEBUG nova.compute.manager [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 854.652990] env[63531]: DEBUG nova.network.neutron [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 854.664110] env[63531]: DEBUG oslo_vmware.api [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5253c480-9f71-71c6-20e7-9576b01d048e, 'name': SearchDatastore_Task, 'duration_secs': 0.013032} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.664884] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.665691] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] 09feb12a-da9f-4bb7-959d-84c5c7c58e99/09feb12a-da9f-4bb7-959d-84c5c7c58e99.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 854.666287] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ad19261a-e564-4b6f-86fa-7734db00dad4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.677483] env[63531]: DEBUG oslo_vmware.api [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Waiting for the task: (returnval){ [ 854.677483] env[63531]: value = "task-1117959" [ 854.677483] env[63531]: _type = "Task" [ 854.677483] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.687913] env[63531]: DEBUG oslo_vmware.api [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': task-1117959, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.696609] env[63531]: DEBUG oslo_vmware.api [None req-dba9d971-47b5-433d-b369-13c4616060ba tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1117957, 'name': CloneVM_Task} progress is 95%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.727573] env[63531]: DEBUG nova.policy [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7c5a958ed3574aff9c1a6a144b42718d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f71c7db0a3614995becc25d9a8dce787', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 854.868183] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Creating Snapshot of the VM instance {{(pid=63531) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 854.868764] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-428ae119-081d-4e7a-bf88-0e5b72a76618 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.881147] env[63531]: DEBUG oslo_vmware.api [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Waiting for the task: (returnval){ [ 854.881147] env[63531]: value = "task-1117960" [ 854.881147] env[63531]: _type = "Task" [ 854.881147] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.889691] env[63531]: DEBUG oslo_vmware.api [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': task-1117960, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.962337] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Releasing lock "refresh_cache-92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.962732] env[63531]: DEBUG nova.compute.manager [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Instance network_info: |[{"id": "b9ea99ca-f03c-4164-b8ae-b68245e39a5a", "address": "fa:16:3e:cf:a3:47", "network": {"id": "421a2d81-9456-47a7-996f-e15b955ef3d5", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-930999152", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.27", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bebde0e281d485ebdd70a1a3b613632", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7a0d5af-5be9-477a-837c-58ef55c717f4", "external-id": "nsx-vlan-transportzone-598", "segmentation_id": 598, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb9ea99ca-f0", "ovs_interfaceid": "b9ea99ca-f03c-4164-b8ae-b68245e39a5a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "cbb5eb21-bc39-4608-820f-a77b3f5fdedf", "address": "fa:16:3e:d0:eb:72", "network": {"id": "4ae7e465-bcda-457c-a060-64abd40b7152", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-251859226", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.59", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "8bebde0e281d485ebdd70a1a3b613632", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1dc3a0d3-9578-4c45-bc16-13ed5b84f5b4", "external-id": "nsx-vlan-transportzone-817", "segmentation_id": 817, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcbb5eb21-bc", "ovs_interfaceid": "cbb5eb21-bc39-4608-820f-a77b3f5fdedf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 854.963122] env[63531]: DEBUG oslo_concurrency.lockutils [req-557089e7-fedb-4711-a820-e43883279ec4 req-f4231da1-4e4e-453a-a990-7eb07c8b97f9 service nova] Acquired lock "refresh_cache-92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.963334] env[63531]: DEBUG nova.network.neutron [req-557089e7-fedb-4711-a820-e43883279ec4 req-f4231da1-4e4e-453a-a990-7eb07c8b97f9 service nova] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Refreshing network info cache for port cbb5eb21-bc39-4608-820f-a77b3f5fdedf {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 854.965165] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cf:a3:47', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e7a0d5af-5be9-477a-837c-58ef55c717f4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b9ea99ca-f03c-4164-b8ae-b68245e39a5a', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:d0:eb:72', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1dc3a0d3-9578-4c45-bc16-13ed5b84f5b4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cbb5eb21-bc39-4608-820f-a77b3f5fdedf', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 854.977860] env[63531]: DEBUG oslo.service.loopingcall [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 854.979612] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 854.979896] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e1c77e5e-0dbc-4132-91d5-43d926c9be20 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.013855] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 855.013855] env[63531]: value = "task-1117961" [ 855.013855] env[63531]: _type = "Task" [ 855.013855] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.026801] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117961, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.071025] env[63531]: DEBUG nova.network.neutron [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Successfully created port: e30711c3-e609-434b-82da-a58252d3678d {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 855.105274] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Releasing lock "refresh_cache-db83adf4-6183-4330-b260-77d1f5daf899" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.105710] env[63531]: DEBUG nova.compute.manager [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: db83adf4-6183-4330-b260-77d1f5daf899] Instance network_info: |[{"id": "895f47c1-21ad-45cc-b1d4-e7a8fa7398cb", "address": "fa:16:3e:07:68:a6", "network": {"id": "887fff70-ea19-458d-a945-e4f3ed6dc185", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2099566288-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "04a4c814245346a9b73253a25a822fef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49a4d142-3f97-47fe-b074-58923c46815e", "external-id": "nsx-vlan-transportzone-565", "segmentation_id": 565, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap895f47c1-21", "ovs_interfaceid": "895f47c1-21ad-45cc-b1d4-e7a8fa7398cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 855.106142] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: db83adf4-6183-4330-b260-77d1f5daf899] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:07:68:a6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '49a4d142-3f97-47fe-b074-58923c46815e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '895f47c1-21ad-45cc-b1d4-e7a8fa7398cb', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 855.117750] env[63531]: DEBUG oslo.service.loopingcall [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 855.118620] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: db83adf4-6183-4330-b260-77d1f5daf899] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 855.118620] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c519db62-826b-40f5-baed-2e9d31229614 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.135859] env[63531]: INFO nova.compute.manager [-] [instance: 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d] Took 1.59 seconds to deallocate network for instance. [ 855.143770] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 855.143770] env[63531]: value = "task-1117962" [ 855.143770] env[63531]: _type = "Task" [ 855.143770] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.153660] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117962, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.157592] env[63531]: DEBUG nova.compute.manager [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 855.191718] env[63531]: DEBUG oslo_vmware.api [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': task-1117959, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.207847] env[63531]: DEBUG oslo_vmware.api [None req-dba9d971-47b5-433d-b369-13c4616060ba tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1117957, 'name': CloneVM_Task} progress is 100%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.396893] env[63531]: DEBUG oslo_vmware.api [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': task-1117960, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.530095] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117961, 'name': CreateVM_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.610018] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1504db78-d82c-4974-951f-2dd6efc1c763 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.616736] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b508802d-5680-455c-a242-7908f569664c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.648528] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1654cc47-4170-4ce7-ab39-6549bb1aa779 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.654650] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdea7036-8b4b-4915-ad5e-81857fd9c28e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.668423] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef8c4852-d58e-4089-936b-262a4c0416ea {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.672700] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117962, 'name': CreateVM_Task, 'duration_secs': 0.498031} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.673600] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: db83adf4-6183-4330-b260-77d1f5daf899] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 855.674221] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8197d52e-21f1-413d-8ef3-59307facd73d" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.674395] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8197d52e-21f1-413d-8ef3-59307facd73d" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.674762] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8197d52e-21f1-413d-8ef3-59307facd73d" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 855.674994] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-295e48e7-bbff-46a5-af93-ec494b966c30 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.684554] env[63531]: DEBUG nova.compute.provider_tree [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 855.692494] env[63531]: DEBUG oslo_vmware.api [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Waiting for the task: (returnval){ [ 855.692494] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5290ac99-6df7-d095-0f75-306891ccc014" [ 855.692494] env[63531]: _type = "Task" [ 855.692494] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.703993] env[63531]: DEBUG oslo_vmware.api [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': task-1117959, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.604092} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.705495] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] 09feb12a-da9f-4bb7-959d-84c5c7c58e99/09feb12a-da9f-4bb7-959d-84c5c7c58e99.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 855.705754] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 855.706529] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-334f1437-0f1b-435c-a919-699f0c65dcb0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.714785] env[63531]: DEBUG oslo_vmware.api [None req-dba9d971-47b5-433d-b369-13c4616060ba tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1117957, 'name': CloneVM_Task, 'duration_secs': 2.195673} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.715435] env[63531]: INFO nova.virt.vmwareapi.vmops [None req-dba9d971-47b5-433d-b369-13c4616060ba tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Created linked-clone VM from snapshot [ 855.715801] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8197d52e-21f1-413d-8ef3-59307facd73d" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.716049] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: db83adf4-6183-4330-b260-77d1f5daf899] Processing image 8197d52e-21f1-413d-8ef3-59307facd73d {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 855.716290] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8197d52e-21f1-413d-8ef3-59307facd73d/8197d52e-21f1-413d-8ef3-59307facd73d.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.716440] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8197d52e-21f1-413d-8ef3-59307facd73d/8197d52e-21f1-413d-8ef3-59307facd73d.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.716621] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 855.717349] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c24ffd92-c376-41f8-8ed4-532438ca23fd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.719953] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b61bd0cb-701a-41ff-9f8a-18c4f2d5e5b8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.722791] env[63531]: DEBUG oslo_vmware.api [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Waiting for the task: (returnval){ [ 855.722791] env[63531]: value = "task-1117963" [ 855.722791] env[63531]: _type = "Task" [ 855.722791] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.729962] env[63531]: DEBUG nova.virt.vmwareapi.images [None req-dba9d971-47b5-433d-b369-13c4616060ba tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Uploading image 2d8829b0-25e0-44ee-a350-e1cd7eb3e470 {{(pid=63531) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 855.733615] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 855.733826] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 855.737020] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-47d40ca2-ea7a-4b29-9ddd-a6771c45d4ff {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.739338] env[63531]: DEBUG oslo_vmware.api [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': task-1117963, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.742167] env[63531]: DEBUG oslo_vmware.api [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Waiting for the task: (returnval){ [ 855.742167] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5253cca7-2582-42eb-c854-e6d13905e91c" [ 855.742167] env[63531]: _type = "Task" [ 855.742167] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.753666] env[63531]: DEBUG oslo_vmware.api [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5253cca7-2582-42eb-c854-e6d13905e91c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.764510] env[63531]: DEBUG oslo_vmware.rw_handles [None req-dba9d971-47b5-433d-b369-13c4616060ba tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 855.764510] env[63531]: value = "vm-244712" [ 855.764510] env[63531]: _type = "VirtualMachine" [ 855.764510] env[63531]: }. {{(pid=63531) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 855.764812] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-5537213e-8a54-470f-8071-40282b085cc7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.772274] env[63531]: DEBUG oslo_vmware.rw_handles [None req-dba9d971-47b5-433d-b369-13c4616060ba tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lease: (returnval){ [ 855.772274] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]526738aa-d56c-84d6-4be4-2ea12c4f9777" [ 855.772274] env[63531]: _type = "HttpNfcLease" [ 855.772274] env[63531]: } obtained for exporting VM: (result){ [ 855.772274] env[63531]: value = "vm-244712" [ 855.772274] env[63531]: _type = "VirtualMachine" [ 855.772274] env[63531]: }. {{(pid=63531) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 855.772274] env[63531]: DEBUG oslo_vmware.api [None req-dba9d971-47b5-433d-b369-13c4616060ba tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the lease: (returnval){ [ 855.772274] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]526738aa-d56c-84d6-4be4-2ea12c4f9777" [ 855.772274] env[63531]: _type = "HttpNfcLease" [ 855.772274] env[63531]: } to be ready. {{(pid=63531) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 855.778288] env[63531]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 855.778288] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]526738aa-d56c-84d6-4be4-2ea12c4f9777" [ 855.778288] env[63531]: _type = "HttpNfcLease" [ 855.778288] env[63531]: } is initializing. {{(pid=63531) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 855.894820] env[63531]: DEBUG oslo_vmware.api [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': task-1117960, 'name': CreateSnapshot_Task, 'duration_secs': 0.955141} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.897560] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Created Snapshot of the VM instance {{(pid=63531) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 855.899287] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7db349bf-c021-4bfa-8feb-0fbc7ef49d28 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.961722] env[63531]: DEBUG nova.network.neutron [req-557089e7-fedb-4711-a820-e43883279ec4 req-f4231da1-4e4e-453a-a990-7eb07c8b97f9 service nova] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Updated VIF entry in instance network info cache for port cbb5eb21-bc39-4608-820f-a77b3f5fdedf. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 855.962163] env[63531]: DEBUG nova.network.neutron [req-557089e7-fedb-4711-a820-e43883279ec4 req-f4231da1-4e4e-453a-a990-7eb07c8b97f9 service nova] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Updating instance_info_cache with network_info: [{"id": "b9ea99ca-f03c-4164-b8ae-b68245e39a5a", "address": "fa:16:3e:cf:a3:47", "network": {"id": "421a2d81-9456-47a7-996f-e15b955ef3d5", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-930999152", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.27", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bebde0e281d485ebdd70a1a3b613632", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7a0d5af-5be9-477a-837c-58ef55c717f4", "external-id": "nsx-vlan-transportzone-598", "segmentation_id": 598, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb9ea99ca-f0", "ovs_interfaceid": "b9ea99ca-f03c-4164-b8ae-b68245e39a5a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "cbb5eb21-bc39-4608-820f-a77b3f5fdedf", "address": "fa:16:3e:d0:eb:72", "network": {"id": "4ae7e465-bcda-457c-a060-64abd40b7152", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-251859226", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.59", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "8bebde0e281d485ebdd70a1a3b613632", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1dc3a0d3-9578-4c45-bc16-13ed5b84f5b4", "external-id": "nsx-vlan-transportzone-817", "segmentation_id": 817, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcbb5eb21-bc", "ovs_interfaceid": "cbb5eb21-bc39-4608-820f-a77b3f5fdedf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.967023] env[63531]: INFO nova.compute.manager [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Rebuilding instance [ 856.015157] env[63531]: DEBUG nova.compute.manager [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 856.016108] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8ad90f7-ede9-4b7e-a4cb-c5318c097498 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.030699] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117961, 'name': CreateVM_Task, 'duration_secs': 0.530652} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.032815] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 856.032815] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.032815] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.032815] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 856.033061] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2ab98967-a787-4146-a9d3-37477b294628 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.038056] env[63531]: DEBUG oslo_vmware.api [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Waiting for the task: (returnval){ [ 856.038056] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52f53ae8-fbae-aa3f-a7a5-314d52b10f66" [ 856.038056] env[63531]: _type = "Task" [ 856.038056] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.046932] env[63531]: DEBUG oslo_vmware.api [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52f53ae8-fbae-aa3f-a7a5-314d52b10f66, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.175961] env[63531]: DEBUG nova.compute.manager [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 856.194237] env[63531]: DEBUG nova.scheduler.client.report [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 856.212701] env[63531]: DEBUG nova.virt.hardware [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 856.212948] env[63531]: DEBUG nova.virt.hardware [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 856.213127] env[63531]: DEBUG nova.virt.hardware [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 856.213360] env[63531]: DEBUG nova.virt.hardware [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 856.213500] env[63531]: DEBUG nova.virt.hardware [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 856.213654] env[63531]: DEBUG nova.virt.hardware [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 856.213866] env[63531]: DEBUG nova.virt.hardware [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 856.214045] env[63531]: DEBUG nova.virt.hardware [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 856.214224] env[63531]: DEBUG nova.virt.hardware [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 856.214391] env[63531]: DEBUG nova.virt.hardware [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 856.214611] env[63531]: DEBUG nova.virt.hardware [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 856.215509] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3147e411-c470-4820-b1ac-559ccc577b50 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.228976] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6371fd0-8e57-4c45-b418-97262c41711f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.250061] env[63531]: DEBUG oslo_vmware.api [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': task-1117963, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071039} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.250391] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 856.251525] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dca1eaf-177c-4a36-81ec-45d4876393b4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.257856] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: db83adf4-6183-4330-b260-77d1f5daf899] Preparing fetch location {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 856.258440] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: db83adf4-6183-4330-b260-77d1f5daf899] Fetch image to [datastore2] OSTACK_IMG_c5b7f03a-1101-4d08-b9a5-cb3358d88051/OSTACK_IMG_c5b7f03a-1101-4d08-b9a5-cb3358d88051.vmdk {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 856.258440] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: db83adf4-6183-4330-b260-77d1f5daf899] Downloading stream optimized image 8197d52e-21f1-413d-8ef3-59307facd73d to [datastore2] OSTACK_IMG_c5b7f03a-1101-4d08-b9a5-cb3358d88051/OSTACK_IMG_c5b7f03a-1101-4d08-b9a5-cb3358d88051.vmdk on the data store datastore2 as vApp {{(pid=63531) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 856.258578] env[63531]: DEBUG nova.virt.vmwareapi.images [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: db83adf4-6183-4330-b260-77d1f5daf899] Downloading image file data 8197d52e-21f1-413d-8ef3-59307facd73d to the ESX as VM named 'OSTACK_IMG_c5b7f03a-1101-4d08-b9a5-cb3358d88051' {{(pid=63531) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 856.276765] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Reconfiguring VM instance instance-00000042 to attach disk [datastore2] 09feb12a-da9f-4bb7-959d-84c5c7c58e99/09feb12a-da9f-4bb7-959d-84c5c7c58e99.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 856.277275] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cedce4a5-d177-4690-b181-16733e6ff633 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.300740] env[63531]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 856.300740] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]526738aa-d56c-84d6-4be4-2ea12c4f9777" [ 856.300740] env[63531]: _type = "HttpNfcLease" [ 856.300740] env[63531]: } is ready. {{(pid=63531) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 856.301106] env[63531]: DEBUG oslo_vmware.rw_handles [None req-dba9d971-47b5-433d-b369-13c4616060ba tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 856.301106] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]526738aa-d56c-84d6-4be4-2ea12c4f9777" [ 856.301106] env[63531]: _type = "HttpNfcLease" [ 856.301106] env[63531]: }. {{(pid=63531) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 856.302873] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78dbfb97-ce68-49d1-8935-9d60e9751c34 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.305492] env[63531]: DEBUG oslo_vmware.api [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Waiting for the task: (returnval){ [ 856.305492] env[63531]: value = "task-1117965" [ 856.305492] env[63531]: _type = "Task" [ 856.305492] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.331635] env[63531]: DEBUG oslo_vmware.rw_handles [None req-dba9d971-47b5-433d-b369-13c4616060ba tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528e465a-6ee9-3b42-7d15-93f9c0e4c59a/disk-0.vmdk from lease info. {{(pid=63531) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 856.331826] env[63531]: DEBUG oslo_vmware.rw_handles [None req-dba9d971-47b5-433d-b369-13c4616060ba tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528e465a-6ee9-3b42-7d15-93f9c0e4c59a/disk-0.vmdk for reading. {{(pid=63531) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 856.397135] env[63531]: DEBUG oslo_vmware.api [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': task-1117965, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.398332] env[63531]: DEBUG nova.compute.manager [req-957f46b8-eda6-4619-8505-d295d6abfe92 req-f762c540-bb0b-45a1-9a4a-901c724582f9 service nova] [instance: 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d] Received event network-vif-deleted-f7815862-8080-412a-8d54-7bad36f1413b {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 856.419136] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Creating linked-clone VM from snapshot {{(pid=63531) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 856.419402] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-60870a89-56af-4dc0-81c8-5e621ddde820 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.427370] env[63531]: DEBUG oslo_vmware.api [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Waiting for the task: (returnval){ [ 856.427370] env[63531]: value = "task-1117966" [ 856.427370] env[63531]: _type = "Task" [ 856.427370] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.435941] env[63531]: DEBUG oslo_vmware.api [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': task-1117966, 'name': CloneVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.436965] env[63531]: DEBUG oslo_vmware.rw_handles [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 856.436965] env[63531]: value = "resgroup-9" [ 856.436965] env[63531]: _type = "ResourcePool" [ 856.436965] env[63531]: }. {{(pid=63531) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 856.437278] env[63531]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-3a690672-5d96-4d42-a583-7a60ecab686f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.458150] env[63531]: DEBUG oslo_vmware.rw_handles [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Lease: (returnval){ [ 856.458150] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52271af5-8b72-fd97-9a53-3baa34477300" [ 856.458150] env[63531]: _type = "HttpNfcLease" [ 856.458150] env[63531]: } obtained for vApp import into resource pool (val){ [ 856.458150] env[63531]: value = "resgroup-9" [ 856.458150] env[63531]: _type = "ResourcePool" [ 856.458150] env[63531]: }. {{(pid=63531) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 856.458150] env[63531]: DEBUG oslo_vmware.api [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Waiting for the lease: (returnval){ [ 856.458150] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52271af5-8b72-fd97-9a53-3baa34477300" [ 856.458150] env[63531]: _type = "HttpNfcLease" [ 856.458150] env[63531]: } to be ready. {{(pid=63531) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 856.465913] env[63531]: DEBUG oslo_concurrency.lockutils [req-557089e7-fedb-4711-a820-e43883279ec4 req-f4231da1-4e4e-453a-a990-7eb07c8b97f9 service nova] Releasing lock "refresh_cache-92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.466237] env[63531]: DEBUG nova.compute.manager [req-557089e7-fedb-4711-a820-e43883279ec4 req-f4231da1-4e4e-453a-a990-7eb07c8b97f9 service nova] [instance: db83adf4-6183-4330-b260-77d1f5daf899] Received event network-changed-895f47c1-21ad-45cc-b1d4-e7a8fa7398cb {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 856.466762] env[63531]: DEBUG nova.compute.manager [req-557089e7-fedb-4711-a820-e43883279ec4 req-f4231da1-4e4e-453a-a990-7eb07c8b97f9 service nova] [instance: db83adf4-6183-4330-b260-77d1f5daf899] Refreshing instance network info cache due to event network-changed-895f47c1-21ad-45cc-b1d4-e7a8fa7398cb. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 856.466841] env[63531]: DEBUG oslo_concurrency.lockutils [req-557089e7-fedb-4711-a820-e43883279ec4 req-f4231da1-4e4e-453a-a990-7eb07c8b97f9 service nova] Acquiring lock "refresh_cache-db83adf4-6183-4330-b260-77d1f5daf899" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.467103] env[63531]: DEBUG oslo_concurrency.lockutils [req-557089e7-fedb-4711-a820-e43883279ec4 req-f4231da1-4e4e-453a-a990-7eb07c8b97f9 service nova] Acquired lock "refresh_cache-db83adf4-6183-4330-b260-77d1f5daf899" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.467366] env[63531]: DEBUG nova.network.neutron [req-557089e7-fedb-4711-a820-e43883279ec4 req-f4231da1-4e4e-453a-a990-7eb07c8b97f9 service nova] [instance: db83adf4-6183-4330-b260-77d1f5daf899] Refreshing network info cache for port 895f47c1-21ad-45cc-b1d4-e7a8fa7398cb {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 856.468495] env[63531]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 856.468495] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52271af5-8b72-fd97-9a53-3baa34477300" [ 856.468495] env[63531]: _type = "HttpNfcLease" [ 856.468495] env[63531]: } is initializing. {{(pid=63531) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 856.510889] env[63531]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-e97de36b-1a2e-4f02-baa1-36381bf199fd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.534052] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 856.537153] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c03d0bec-5c6b-4d63-8e47-9ee7514b9c87 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.552594] env[63531]: DEBUG oslo_vmware.api [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 856.552594] env[63531]: value = "task-1117968" [ 856.552594] env[63531]: _type = "Task" [ 856.552594] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.555631] env[63531]: DEBUG oslo_vmware.api [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52f53ae8-fbae-aa3f-a7a5-314d52b10f66, 'name': SearchDatastore_Task, 'duration_secs': 0.054317} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.559109] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.559369] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 856.559604] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.559752] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.559919] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 856.560201] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d5560c69-657b-44a5-9406-0364b7eda0b9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.567928] env[63531]: DEBUG oslo_vmware.api [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1117968, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.569219] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 856.569393] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 856.570123] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cde64775-1585-4198-9a04-dfc509b9155c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.576480] env[63531]: DEBUG oslo_vmware.api [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Waiting for the task: (returnval){ [ 856.576480] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52fed032-c593-b11d-2a7e-9b867a61cc6e" [ 856.576480] env[63531]: _type = "Task" [ 856.576480] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.584534] env[63531]: DEBUG oslo_vmware.api [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52fed032-c593-b11d-2a7e-9b867a61cc6e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.706570] env[63531]: DEBUG nova.compute.manager [req-874c4fc9-df88-45cc-935f-8c74fbe00c06 req-b3c7cda7-0b10-4f5e-9ca6-d0f21bf731be service nova] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Received event network-vif-plugged-e30711c3-e609-434b-82da-a58252d3678d {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 856.707517] env[63531]: DEBUG oslo_concurrency.lockutils [req-874c4fc9-df88-45cc-935f-8c74fbe00c06 req-b3c7cda7-0b10-4f5e-9ca6-d0f21bf731be service nova] Acquiring lock "4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.707722] env[63531]: DEBUG oslo_concurrency.lockutils [req-874c4fc9-df88-45cc-935f-8c74fbe00c06 req-b3c7cda7-0b10-4f5e-9ca6-d0f21bf731be service nova] Lock "4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.708044] env[63531]: DEBUG oslo_concurrency.lockutils [req-874c4fc9-df88-45cc-935f-8c74fbe00c06 req-b3c7cda7-0b10-4f5e-9ca6-d0f21bf731be service nova] Lock "4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.708452] env[63531]: DEBUG nova.compute.manager [req-874c4fc9-df88-45cc-935f-8c74fbe00c06 req-b3c7cda7-0b10-4f5e-9ca6-d0f21bf731be service nova] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] No waiting events found dispatching network-vif-plugged-e30711c3-e609-434b-82da-a58252d3678d {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 856.708745] env[63531]: WARNING nova.compute.manager [req-874c4fc9-df88-45cc-935f-8c74fbe00c06 req-b3c7cda7-0b10-4f5e-9ca6-d0f21bf731be service nova] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Received unexpected event network-vif-plugged-e30711c3-e609-434b-82da-a58252d3678d for instance with vm_state building and task_state spawning. [ 856.709665] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.576s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.710241] env[63531]: DEBUG nova.compute.manager [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 856.716898] env[63531]: DEBUG oslo_concurrency.lockutils [None req-932f1c32-d337-45aa-af0e-9ce7ab65093a tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.778s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.716898] env[63531]: DEBUG nova.objects.instance [None req-932f1c32-d337-45aa-af0e-9ce7ab65093a tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Lazy-loading 'resources' on Instance uuid 5e9042b3-4dc2-4fa3-a664-c4b49a22e400 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 856.833460] env[63531]: DEBUG oslo_vmware.api [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': task-1117965, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.939336] env[63531]: DEBUG oslo_vmware.api [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': task-1117966, 'name': CloneVM_Task} progress is 94%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.965417] env[63531]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 856.965417] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52271af5-8b72-fd97-9a53-3baa34477300" [ 856.965417] env[63531]: _type = "HttpNfcLease" [ 856.965417] env[63531]: } is initializing. {{(pid=63531) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 857.070122] env[63531]: DEBUG oslo_vmware.api [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1117968, 'name': PowerOffVM_Task, 'duration_secs': 0.380791} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.070690] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 857.070907] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 857.071739] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02b542be-3ffd-49f9-8514-9af1802e0398 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.082064] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 857.082858] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9812cd90-377f-455a-8ff5-175ffd9b9d59 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.088600] env[63531]: DEBUG oslo_vmware.api [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52fed032-c593-b11d-2a7e-9b867a61cc6e, 'name': SearchDatastore_Task, 'duration_secs': 0.011657} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.090205] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-38f21cd7-9b35-4db4-b143-11179ee5ed80 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.097026] env[63531]: DEBUG oslo_vmware.api [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Waiting for the task: (returnval){ [ 857.097026] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52af45af-e412-45af-733f-5a99956816ae" [ 857.097026] env[63531]: _type = "Task" [ 857.097026] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.108783] env[63531]: DEBUG oslo_vmware.api [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52af45af-e412-45af-733f-5a99956816ae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.160753] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 857.161451] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Deleting contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 857.161771] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Deleting the datastore file [datastore2] 2b0e4285-310a-439c-bc50-000a7c5ef7f9 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 857.162393] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-160cb654-d1ad-4554-bc39-c7ca99a97bfb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.169488] env[63531]: DEBUG oslo_vmware.api [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 857.169488] env[63531]: value = "task-1117970" [ 857.169488] env[63531]: _type = "Task" [ 857.169488] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.179080] env[63531]: DEBUG oslo_vmware.api [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1117970, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.218023] env[63531]: DEBUG nova.compute.utils [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 857.218023] env[63531]: DEBUG nova.compute.manager [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 857.218023] env[63531]: DEBUG nova.network.neutron [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 857.336425] env[63531]: DEBUG oslo_vmware.api [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': task-1117965, 'name': ReconfigVM_Task, 'duration_secs': 0.604173} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.338224] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Reconfigured VM instance instance-00000042 to attach disk [datastore2] 09feb12a-da9f-4bb7-959d-84c5c7c58e99/09feb12a-da9f-4bb7-959d-84c5c7c58e99.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 857.338764] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5c4caf71-b61a-4f42-833e-8cad7773093f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.350156] env[63531]: DEBUG oslo_vmware.api [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Waiting for the task: (returnval){ [ 857.350156] env[63531]: value = "task-1117971" [ 857.350156] env[63531]: _type = "Task" [ 857.350156] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.367441] env[63531]: DEBUG oslo_vmware.api [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': task-1117971, 'name': Rename_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.373848] env[63531]: DEBUG nova.network.neutron [req-557089e7-fedb-4711-a820-e43883279ec4 req-f4231da1-4e4e-453a-a990-7eb07c8b97f9 service nova] [instance: db83adf4-6183-4330-b260-77d1f5daf899] Updated VIF entry in instance network info cache for port 895f47c1-21ad-45cc-b1d4-e7a8fa7398cb. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 857.373848] env[63531]: DEBUG nova.network.neutron [req-557089e7-fedb-4711-a820-e43883279ec4 req-f4231da1-4e4e-453a-a990-7eb07c8b97f9 service nova] [instance: db83adf4-6183-4330-b260-77d1f5daf899] Updating instance_info_cache with network_info: [{"id": "895f47c1-21ad-45cc-b1d4-e7a8fa7398cb", "address": "fa:16:3e:07:68:a6", "network": {"id": "887fff70-ea19-458d-a945-e4f3ed6dc185", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2099566288-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "04a4c814245346a9b73253a25a822fef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49a4d142-3f97-47fe-b074-58923c46815e", "external-id": "nsx-vlan-transportzone-565", "segmentation_id": 565, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap895f47c1-21", "ovs_interfaceid": "895f47c1-21ad-45cc-b1d4-e7a8fa7398cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 857.441077] env[63531]: DEBUG oslo_vmware.api [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': task-1117966, 'name': CloneVM_Task} progress is 100%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.466762] env[63531]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 857.466762] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52271af5-8b72-fd97-9a53-3baa34477300" [ 857.466762] env[63531]: _type = "HttpNfcLease" [ 857.466762] env[63531]: } is ready. {{(pid=63531) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 857.467485] env[63531]: DEBUG oslo_vmware.rw_handles [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 857.467485] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52271af5-8b72-fd97-9a53-3baa34477300" [ 857.467485] env[63531]: _type = "HttpNfcLease" [ 857.467485] env[63531]: }. {{(pid=63531) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 857.468281] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33c62fff-5a85-4668-9216-a4c44a1c7025 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.480929] env[63531]: DEBUG oslo_vmware.rw_handles [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5235b400-ae02-40d0-2d99-d227e76ab482/disk-0.vmdk from lease info. {{(pid=63531) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 857.481254] env[63531]: DEBUG oslo_vmware.rw_handles [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5235b400-ae02-40d0-2d99-d227e76ab482/disk-0.vmdk. {{(pid=63531) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 857.541627] env[63531]: DEBUG nova.network.neutron [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Successfully updated port: e30711c3-e609-434b-82da-a58252d3678d {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 857.544041] env[63531]: DEBUG nova.policy [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7c5a958ed3574aff9c1a6a144b42718d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f71c7db0a3614995becc25d9a8dce787', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 857.550741] env[63531]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-ff33d89e-efad-4570-8963-d1146ff946e2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.610500] env[63531]: DEBUG oslo_vmware.api [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52af45af-e412-45af-733f-5a99956816ae, 'name': SearchDatastore_Task, 'duration_secs': 0.011848} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.610500] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.610710] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b/92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 857.610935] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-533f7f50-0cad-4389-884a-b7719ba37868 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.617861] env[63531]: DEBUG oslo_vmware.api [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Waiting for the task: (returnval){ [ 857.617861] env[63531]: value = "task-1117972" [ 857.617861] env[63531]: _type = "Task" [ 857.617861] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.629260] env[63531]: DEBUG oslo_vmware.api [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Task: {'id': task-1117972, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.680294] env[63531]: DEBUG oslo_vmware.api [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1117970, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.21972} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.683524] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 857.683817] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Deleted contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 857.684099] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 857.701054] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8a579f8-4d4c-4338-8e47-4f14197cc848 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.708854] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba2b7a44-b88e-4075-abe8-b9fd1d3f4a72 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.740743] env[63531]: DEBUG nova.compute.manager [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 857.745290] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a886d1c4-b794-47e9-94ae-ce00044f93c7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.754206] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0594b621-d390-4601-97ec-4f98fe1bba31 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.769263] env[63531]: DEBUG nova.compute.provider_tree [None req-932f1c32-d337-45aa-af0e-9ce7ab65093a tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 857.839330] env[63531]: DEBUG nova.network.neutron [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Successfully created port: d5d578b0-1f4a-4a46-8fac-1e625d48c2bc {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 857.865677] env[63531]: DEBUG oslo_vmware.api [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': task-1117971, 'name': Rename_Task, 'duration_secs': 0.244051} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.866239] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 857.866760] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fbaa151d-5b09-4d1e-9111-2f897b6dec13 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.876537] env[63531]: DEBUG oslo_vmware.api [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Waiting for the task: (returnval){ [ 857.876537] env[63531]: value = "task-1117973" [ 857.876537] env[63531]: _type = "Task" [ 857.876537] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.881546] env[63531]: DEBUG oslo_concurrency.lockutils [req-557089e7-fedb-4711-a820-e43883279ec4 req-f4231da1-4e4e-453a-a990-7eb07c8b97f9 service nova] Releasing lock "refresh_cache-db83adf4-6183-4330-b260-77d1f5daf899" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.887940] env[63531]: DEBUG oslo_vmware.api [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': task-1117973, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.942631] env[63531]: DEBUG oslo_vmware.api [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': task-1117966, 'name': CloneVM_Task, 'duration_secs': 1.015246} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.943222] env[63531]: INFO nova.virt.vmwareapi.vmops [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Created linked-clone VM from snapshot [ 857.944334] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc70f448-f6fb-49f0-97e7-4f7aa5dc5baa {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.954282] env[63531]: DEBUG nova.virt.vmwareapi.images [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Uploading image d477826e-3e39-4ca7-bc24-998780166e1c {{(pid=63531) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 858.032224] env[63531]: DEBUG oslo_vmware.rw_handles [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 858.032224] env[63531]: value = "vm-244716" [ 858.032224] env[63531]: _type = "VirtualMachine" [ 858.032224] env[63531]: }. {{(pid=63531) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 858.032789] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-29af7d42-e8d3-4332-8f56-d5c424f37f1d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.040910] env[63531]: DEBUG oslo_vmware.rw_handles [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Lease: (returnval){ [ 858.040910] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]520656a5-361f-4488-dce3-8c0bb914cb8c" [ 858.040910] env[63531]: _type = "HttpNfcLease" [ 858.040910] env[63531]: } obtained for exporting VM: (result){ [ 858.040910] env[63531]: value = "vm-244716" [ 858.040910] env[63531]: _type = "VirtualMachine" [ 858.040910] env[63531]: }. {{(pid=63531) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 858.041246] env[63531]: DEBUG oslo_vmware.api [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Waiting for the lease: (returnval){ [ 858.041246] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]520656a5-361f-4488-dce3-8c0bb914cb8c" [ 858.041246] env[63531]: _type = "HttpNfcLease" [ 858.041246] env[63531]: } to be ready. {{(pid=63531) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 858.047921] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquiring lock "refresh_cache-4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.047921] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquired lock "refresh_cache-4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.047921] env[63531]: DEBUG nova.network.neutron [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 858.054164] env[63531]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 858.054164] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]520656a5-361f-4488-dce3-8c0bb914cb8c" [ 858.054164] env[63531]: _type = "HttpNfcLease" [ 858.054164] env[63531]: } is initializing. {{(pid=63531) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 858.134321] env[63531]: DEBUG oslo_vmware.api [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Task: {'id': task-1117972, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.276225] env[63531]: DEBUG nova.scheduler.client.report [None req-932f1c32-d337-45aa-af0e-9ce7ab65093a tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 858.390344] env[63531]: DEBUG oslo_vmware.api [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': task-1117973, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.444019] env[63531]: DEBUG oslo_vmware.rw_handles [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Completed reading data from the image iterator. {{(pid=63531) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 858.444019] env[63531]: DEBUG oslo_vmware.rw_handles [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5235b400-ae02-40d0-2d99-d227e76ab482/disk-0.vmdk. {{(pid=63531) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 858.444019] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e0e5ae0-777d-463b-af60-26a77bac7b7f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.450521] env[63531]: DEBUG oslo_vmware.rw_handles [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5235b400-ae02-40d0-2d99-d227e76ab482/disk-0.vmdk is in state: ready. {{(pid=63531) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 858.450894] env[63531]: DEBUG oslo_vmware.rw_handles [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5235b400-ae02-40d0-2d99-d227e76ab482/disk-0.vmdk. {{(pid=63531) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 858.451232] env[63531]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-eeae3acc-c719-44cc-a4e3-9667e8fef08c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.553802] env[63531]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 858.553802] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]520656a5-361f-4488-dce3-8c0bb914cb8c" [ 858.553802] env[63531]: _type = "HttpNfcLease" [ 858.553802] env[63531]: } is ready. {{(pid=63531) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 858.554297] env[63531]: DEBUG oslo_vmware.rw_handles [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 858.554297] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]520656a5-361f-4488-dce3-8c0bb914cb8c" [ 858.554297] env[63531]: _type = "HttpNfcLease" [ 858.554297] env[63531]: }. {{(pid=63531) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 858.555967] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c560087-d363-4758-a5fb-d55620d39d8e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.565911] env[63531]: DEBUG oslo_vmware.rw_handles [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52a2690c-2c06-2f74-e0fe-59cf659a8252/disk-0.vmdk from lease info. {{(pid=63531) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 858.566644] env[63531]: DEBUG oslo_vmware.rw_handles [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52a2690c-2c06-2f74-e0fe-59cf659a8252/disk-0.vmdk for reading. {{(pid=63531) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 858.632380] env[63531]: DEBUG nova.network.neutron [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 858.641203] env[63531]: DEBUG nova.compute.manager [req-bacbf2f2-5b7b-43b8-9157-f92d6e9d6351 req-9b872bcc-d4ef-498e-b3a5-be12d3b9a312 service nova] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Received event network-changed-e30711c3-e609-434b-82da-a58252d3678d {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 858.644666] env[63531]: DEBUG nova.compute.manager [req-bacbf2f2-5b7b-43b8-9157-f92d6e9d6351 req-9b872bcc-d4ef-498e-b3a5-be12d3b9a312 service nova] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Refreshing instance network info cache due to event network-changed-e30711c3-e609-434b-82da-a58252d3678d. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 858.644666] env[63531]: DEBUG oslo_concurrency.lockutils [req-bacbf2f2-5b7b-43b8-9157-f92d6e9d6351 req-9b872bcc-d4ef-498e-b3a5-be12d3b9a312 service nova] Acquiring lock "refresh_cache-4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.651328] env[63531]: DEBUG oslo_vmware.api [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Task: {'id': task-1117972, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.528535} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.652099] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b/92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 858.652226] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 858.652625] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ef62ca87-0fdc-4ffe-b8f0-cc6ebc327c78 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.661124] env[63531]: DEBUG oslo_vmware.api [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Waiting for the task: (returnval){ [ 858.661124] env[63531]: value = "task-1117975" [ 858.661124] env[63531]: _type = "Task" [ 858.661124] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.673103] env[63531]: DEBUG oslo_vmware.api [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Task: {'id': task-1117975, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.674591] env[63531]: DEBUG oslo_vmware.rw_handles [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5235b400-ae02-40d0-2d99-d227e76ab482/disk-0.vmdk. {{(pid=63531) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 858.674876] env[63531]: INFO nova.virt.vmwareapi.images [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: db83adf4-6183-4330-b260-77d1f5daf899] Downloaded image file data 8197d52e-21f1-413d-8ef3-59307facd73d [ 858.675723] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff4e0bb7-048f-4a20-8a37-89824a37b800 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.699064] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-377d9056-96bf-48fe-bc00-8831acd03daa {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.703534] env[63531]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-8ce00b23-c549-47bb-9ee1-1ff73dec6883 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.743135] env[63531]: DEBUG nova.virt.hardware [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 858.744118] env[63531]: DEBUG nova.virt.hardware [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 858.744294] env[63531]: DEBUG nova.virt.hardware [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 858.744508] env[63531]: DEBUG nova.virt.hardware [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 858.744722] env[63531]: DEBUG nova.virt.hardware [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 858.744936] env[63531]: DEBUG nova.virt.hardware [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 858.745271] env[63531]: DEBUG nova.virt.hardware [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 858.745551] env[63531]: DEBUG nova.virt.hardware [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 858.745814] env[63531]: DEBUG nova.virt.hardware [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 858.745996] env[63531]: DEBUG nova.virt.hardware [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 858.746277] env[63531]: DEBUG nova.virt.hardware [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 858.746990] env[63531]: INFO nova.virt.vmwareapi.images [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: db83adf4-6183-4330-b260-77d1f5daf899] The imported VM was unregistered [ 858.748910] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: db83adf4-6183-4330-b260-77d1f5daf899] Caching image {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 858.749199] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Creating directory with path [datastore2] devstack-image-cache_base/8197d52e-21f1-413d-8ef3-59307facd73d {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 858.751663] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f64abf45-7581-4a35-867a-178851adacf5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.756111] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6bbef34e-ccb7-4881-b570-76540f7a75bf {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.758962] env[63531]: DEBUG nova.compute.manager [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 858.771736] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d357bea3-a816-44c2-8e1b-05f3c0a61f16 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.781159] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Created directory with path [datastore2] devstack-image-cache_base/8197d52e-21f1-413d-8ef3-59307facd73d {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 858.781159] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_c5b7f03a-1101-4d08-b9a5-cb3358d88051/OSTACK_IMG_c5b7f03a-1101-4d08-b9a5-cb3358d88051.vmdk to [datastore2] devstack-image-cache_base/8197d52e-21f1-413d-8ef3-59307facd73d/8197d52e-21f1-413d-8ef3-59307facd73d.vmdk. {{(pid=63531) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 858.781159] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-ed899d70-357b-4687-af58-33bd0ab99b02 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.786748] env[63531]: DEBUG oslo_concurrency.lockutils [None req-932f1c32-d337-45aa-af0e-9ce7ab65093a tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.073s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.800346] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ead7921a-cb8c-490f-a3ab-414729b031bc tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.072s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.801309] env[63531]: DEBUG nova.objects.instance [None req-ead7921a-cb8c-490f-a3ab-414729b031bc tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Lazy-loading 'resources' on Instance uuid 0197e03d-6c36-4e73-9472-8e3319eb89ce {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 858.803267] env[63531]: DEBUG oslo_vmware.api [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Waiting for the task: (returnval){ [ 858.803267] env[63531]: value = "task-1117977" [ 858.803267] env[63531]: _type = "Task" [ 858.803267] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.803898] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2a:0b:33', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bc6e6fe1-c4f5-4389-a49f-0978060eebb4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '13365dfb-f81c-4e56-bf30-0a09dec4f561', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 858.812691] env[63531]: DEBUG oslo.service.loopingcall [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 858.816552] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 858.820580] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-21d84e78-9be7-4281-ade0-a832e3b751ee {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.836979] env[63531]: INFO nova.scheduler.client.report [None req-932f1c32-d337-45aa-af0e-9ce7ab65093a tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Deleted allocations for instance 5e9042b3-4dc2-4fa3-a664-c4b49a22e400 [ 858.849070] env[63531]: DEBUG oslo_vmware.api [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1117977, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.851372] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 858.851372] env[63531]: value = "task-1117978" [ 858.851372] env[63531]: _type = "Task" [ 858.851372] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.859912] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117978, 'name': CreateVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.863788] env[63531]: DEBUG nova.network.neutron [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Updating instance_info_cache with network_info: [{"id": "e30711c3-e609-434b-82da-a58252d3678d", "address": "fa:16:3e:54:c1:b7", "network": {"id": "913045d9-e90b-4366-987c-7706b73a6147", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1696510389-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f71c7db0a3614995becc25d9a8dce787", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c297fe21-cd0b-4226-813b-a65d2358d034", "external-id": "nsx-vlan-transportzone-98", "segmentation_id": 98, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape30711c3-e6", "ovs_interfaceid": "e30711c3-e609-434b-82da-a58252d3678d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 858.891332] env[63531]: DEBUG oslo_vmware.api [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': task-1117973, 'name': PowerOnVM_Task, 'duration_secs': 0.656096} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.892127] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 858.892127] env[63531]: DEBUG nova.compute.manager [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 858.892993] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84a54bc7-5150-424a-904d-e174589d87d0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.954168] env[63531]: DEBUG nova.virt.hardware [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 858.954465] env[63531]: DEBUG nova.virt.hardware [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 858.955336] env[63531]: DEBUG nova.virt.hardware [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 858.955336] env[63531]: DEBUG nova.virt.hardware [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 858.955336] env[63531]: DEBUG nova.virt.hardware [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 858.955336] env[63531]: DEBUG nova.virt.hardware [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 858.955579] env[63531]: DEBUG nova.virt.hardware [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 858.955624] env[63531]: DEBUG nova.virt.hardware [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 858.955754] env[63531]: DEBUG nova.virt.hardware [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 858.955919] env[63531]: DEBUG nova.virt.hardware [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 858.956107] env[63531]: DEBUG nova.virt.hardware [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 858.957210] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d15d61d-fa2f-4760-a4f6-6f0f402d1459 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.966875] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31b150e8-68af-40c0-9afd-7987f0779853 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.173511] env[63531]: DEBUG oslo_vmware.api [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Task: {'id': task-1117975, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.18823} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.174036] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 859.174917] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d12c2050-ac46-4f0d-8804-96abb1690d1a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.203217] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Reconfiguring VM instance instance-00000044 to attach disk [datastore2] 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b/92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 859.206208] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-41128f17-e648-4e96-bd33-514727f892e1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.227727] env[63531]: DEBUG oslo_vmware.api [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Waiting for the task: (returnval){ [ 859.227727] env[63531]: value = "task-1117979" [ 859.227727] env[63531]: _type = "Task" [ 859.227727] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.239329] env[63531]: DEBUG oslo_vmware.api [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Task: {'id': task-1117979, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.256111] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-022d6883-8d4b-4fbb-a268-f4c7e06eaf98 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.264613] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5339d0e-2414-4e9f-9880-39966b4b53d1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.300119] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df8b5017-1875-4c35-983c-01355a88a5d7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.313873] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b4452f7-c109-4a68-a43d-21e3a9108f98 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.339762] env[63531]: DEBUG nova.compute.provider_tree [None req-ead7921a-cb8c-490f-a3ab-414729b031bc tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 859.347690] env[63531]: DEBUG oslo_vmware.api [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1117977, 'name': MoveVirtualDisk_Task} progress is 9%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.348441] env[63531]: DEBUG oslo_concurrency.lockutils [None req-932f1c32-d337-45aa-af0e-9ce7ab65093a tempest-MigrationsAdminTest-242187684 tempest-MigrationsAdminTest-242187684-project-member] Lock "5e9042b3-4dc2-4fa3-a664-c4b49a22e400" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.342s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.366124] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Releasing lock "refresh_cache-4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.366529] env[63531]: DEBUG nova.compute.manager [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Instance network_info: |[{"id": "e30711c3-e609-434b-82da-a58252d3678d", "address": "fa:16:3e:54:c1:b7", "network": {"id": "913045d9-e90b-4366-987c-7706b73a6147", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1696510389-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f71c7db0a3614995becc25d9a8dce787", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c297fe21-cd0b-4226-813b-a65d2358d034", "external-id": "nsx-vlan-transportzone-98", "segmentation_id": 98, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape30711c3-e6", "ovs_interfaceid": "e30711c3-e609-434b-82da-a58252d3678d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 859.367353] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117978, 'name': CreateVM_Task} progress is 25%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.367809] env[63531]: DEBUG oslo_concurrency.lockutils [req-bacbf2f2-5b7b-43b8-9157-f92d6e9d6351 req-9b872bcc-d4ef-498e-b3a5-be12d3b9a312 service nova] Acquired lock "refresh_cache-4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.368039] env[63531]: DEBUG nova.network.neutron [req-bacbf2f2-5b7b-43b8-9157-f92d6e9d6351 req-9b872bcc-d4ef-498e-b3a5-be12d3b9a312 service nova] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Refreshing network info cache for port e30711c3-e609-434b-82da-a58252d3678d {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 859.369817] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:54:c1:b7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c297fe21-cd0b-4226-813b-a65d2358d034', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e30711c3-e609-434b-82da-a58252d3678d', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 859.378395] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Creating folder: Project (f71c7db0a3614995becc25d9a8dce787). Parent ref: group-v244585. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 859.380879] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9b2df3ea-e55d-4748-b092-9de06b608189 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.394184] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Created folder: Project (f71c7db0a3614995becc25d9a8dce787) in parent group-v244585. [ 859.394419] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Creating folder: Instances. Parent ref: group-v244719. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 859.394719] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3f5df7a7-7b8f-4d99-be7e-6fa4eb2cc7c3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.406153] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Created folder: Instances in parent group-v244719. [ 859.406467] env[63531]: DEBUG oslo.service.loopingcall [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 859.406686] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 859.411636] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b60fa054-d58f-4d65-867c-d256d6877e27 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.438342] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.441222] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 859.441222] env[63531]: value = "task-1117982" [ 859.441222] env[63531]: _type = "Task" [ 859.441222] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.451271] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117982, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.503883] env[63531]: DEBUG nova.network.neutron [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Successfully updated port: d5d578b0-1f4a-4a46-8fac-1e625d48c2bc {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 859.738808] env[63531]: DEBUG oslo_vmware.api [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Task: {'id': task-1117979, 'name': ReconfigVM_Task, 'duration_secs': 0.502021} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.739159] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Reconfigured VM instance instance-00000044 to attach disk [datastore2] 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b/92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 859.740050] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ec42f30d-86ac-404d-9e55-16982561598a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.746568] env[63531]: DEBUG oslo_vmware.api [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Waiting for the task: (returnval){ [ 859.746568] env[63531]: value = "task-1117983" [ 859.746568] env[63531]: _type = "Task" [ 859.746568] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.755946] env[63531]: DEBUG oslo_vmware.api [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Task: {'id': task-1117983, 'name': Rename_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.832232] env[63531]: DEBUG oslo_vmware.api [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1117977, 'name': MoveVirtualDisk_Task} progress is 32%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.851736] env[63531]: DEBUG nova.scheduler.client.report [None req-ead7921a-cb8c-490f-a3ab-414729b031bc tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 859.866307] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117978, 'name': CreateVM_Task, 'duration_secs': 0.66767} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.866464] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 859.867681] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.867893] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.868694] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 859.868694] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4665d96b-1dcf-467f-90ad-fd1f39d274b3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.876159] env[63531]: DEBUG oslo_vmware.api [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 859.876159] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]528b6b36-e26d-5f2a-71b2-9636d00beec3" [ 859.876159] env[63531]: _type = "Task" [ 859.876159] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.891278] env[63531]: DEBUG oslo_vmware.api [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]528b6b36-e26d-5f2a-71b2-9636d00beec3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.954060] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117982, 'name': CreateVM_Task, 'duration_secs': 0.419863} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.954315] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 859.955356] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.955689] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.956346] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 859.956715] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-27460252-228c-4538-92ce-b859e6189afd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.962566] env[63531]: DEBUG oslo_vmware.api [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 859.962566] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]521e69f1-38f6-0eb2-0619-63f2fea14440" [ 859.962566] env[63531]: _type = "Task" [ 859.962566] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.972373] env[63531]: DEBUG oslo_vmware.api [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]521e69f1-38f6-0eb2-0619-63f2fea14440, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.005457] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquiring lock "refresh_cache-879924d4-b465-4102-a0e3-c7b2be7ef08b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.005886] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquired lock "refresh_cache-879924d4-b465-4102-a0e3-c7b2be7ef08b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.006435] env[63531]: DEBUG nova.network.neutron [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 860.152230] env[63531]: DEBUG nova.network.neutron [req-bacbf2f2-5b7b-43b8-9157-f92d6e9d6351 req-9b872bcc-d4ef-498e-b3a5-be12d3b9a312 service nova] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Updated VIF entry in instance network info cache for port e30711c3-e609-434b-82da-a58252d3678d. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 860.152693] env[63531]: DEBUG nova.network.neutron [req-bacbf2f2-5b7b-43b8-9157-f92d6e9d6351 req-9b872bcc-d4ef-498e-b3a5-be12d3b9a312 service nova] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Updating instance_info_cache with network_info: [{"id": "e30711c3-e609-434b-82da-a58252d3678d", "address": "fa:16:3e:54:c1:b7", "network": {"id": "913045d9-e90b-4366-987c-7706b73a6147", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1696510389-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f71c7db0a3614995becc25d9a8dce787", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c297fe21-cd0b-4226-813b-a65d2358d034", "external-id": "nsx-vlan-transportzone-98", "segmentation_id": 98, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape30711c3-e6", "ovs_interfaceid": "e30711c3-e609-434b-82da-a58252d3678d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.259136] env[63531]: DEBUG oslo_vmware.api [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Task: {'id': task-1117983, 'name': Rename_Task, 'duration_secs': 0.217403} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.259476] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 860.259785] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-19cc7134-dc13-4af8-8770-b8a3ee8b509b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.267386] env[63531]: DEBUG oslo_vmware.api [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Waiting for the task: (returnval){ [ 860.267386] env[63531]: value = "task-1117984" [ 860.267386] env[63531]: _type = "Task" [ 860.267386] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.277142] env[63531]: DEBUG oslo_vmware.api [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Task: {'id': task-1117984, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.338942] env[63531]: DEBUG oslo_vmware.api [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1117977, 'name': MoveVirtualDisk_Task} progress is 52%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.357339] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ead7921a-cb8c-490f-a3ab-414729b031bc tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.557s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.360252] env[63531]: DEBUG oslo_concurrency.lockutils [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.556s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.362098] env[63531]: INFO nova.compute.claims [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 860.391140] env[63531]: DEBUG oslo_vmware.api [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]528b6b36-e26d-5f2a-71b2-9636d00beec3, 'name': SearchDatastore_Task, 'duration_secs': 0.078124} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.391761] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.391761] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 860.392085] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.392254] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.392449] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 860.392744] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-39037ca5-36af-4135-8deb-d5c908506bc8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.402411] env[63531]: INFO nova.scheduler.client.report [None req-ead7921a-cb8c-490f-a3ab-414729b031bc tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Deleted allocations for instance 0197e03d-6c36-4e73-9472-8e3319eb89ce [ 860.411611] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 860.411611] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 860.411611] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-99346b69-c86a-4d1a-b5f9-f5fe7f15a3b3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.418438] env[63531]: DEBUG oslo_vmware.api [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 860.418438] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52543ccc-6756-ee12-e397-287ff06a953a" [ 860.418438] env[63531]: _type = "Task" [ 860.418438] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.430188] env[63531]: DEBUG oslo_vmware.api [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52543ccc-6756-ee12-e397-287ff06a953a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.475854] env[63531]: DEBUG oslo_vmware.api [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]521e69f1-38f6-0eb2-0619-63f2fea14440, 'name': SearchDatastore_Task, 'duration_secs': 0.018956} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.476268] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.476629] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 860.476871] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.477071] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.477292] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 860.478068] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b5d08feb-b53c-4955-aa1a-4ff35ad755c1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.488799] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 860.489022] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 860.490137] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ea82bbd8-47b3-4f82-bc44-fe31ab99b777 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.496477] env[63531]: DEBUG oslo_vmware.api [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 860.496477] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52a57765-4084-2ee7-21b4-fe2bc9b654ef" [ 860.496477] env[63531]: _type = "Task" [ 860.496477] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.507571] env[63531]: DEBUG oslo_vmware.api [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52a57765-4084-2ee7-21b4-fe2bc9b654ef, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.544285] env[63531]: DEBUG nova.network.neutron [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 860.629188] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1b340bf7-08de-4d21-a861-c6e587976b10 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Acquiring lock "09feb12a-da9f-4bb7-959d-84c5c7c58e99" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.629481] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1b340bf7-08de-4d21-a861-c6e587976b10 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Lock "09feb12a-da9f-4bb7-959d-84c5c7c58e99" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.629826] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1b340bf7-08de-4d21-a861-c6e587976b10 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Acquiring lock "09feb12a-da9f-4bb7-959d-84c5c7c58e99-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.630066] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1b340bf7-08de-4d21-a861-c6e587976b10 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Lock "09feb12a-da9f-4bb7-959d-84c5c7c58e99-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.630308] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1b340bf7-08de-4d21-a861-c6e587976b10 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Lock "09feb12a-da9f-4bb7-959d-84c5c7c58e99-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.635973] env[63531]: INFO nova.compute.manager [None req-1b340bf7-08de-4d21-a861-c6e587976b10 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Terminating instance [ 860.638669] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1b340bf7-08de-4d21-a861-c6e587976b10 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Acquiring lock "refresh_cache-09feb12a-da9f-4bb7-959d-84c5c7c58e99" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.638812] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1b340bf7-08de-4d21-a861-c6e587976b10 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Acquired lock "refresh_cache-09feb12a-da9f-4bb7-959d-84c5c7c58e99" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.639082] env[63531]: DEBUG nova.network.neutron [None req-1b340bf7-08de-4d21-a861-c6e587976b10 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 860.655593] env[63531]: DEBUG oslo_concurrency.lockutils [req-bacbf2f2-5b7b-43b8-9157-f92d6e9d6351 req-9b872bcc-d4ef-498e-b3a5-be12d3b9a312 service nova] Releasing lock "refresh_cache-4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.723472] env[63531]: DEBUG nova.network.neutron [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Updating instance_info_cache with network_info: [{"id": "d5d578b0-1f4a-4a46-8fac-1e625d48c2bc", "address": "fa:16:3e:ca:1a:aa", "network": {"id": "913045d9-e90b-4366-987c-7706b73a6147", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1696510389-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f71c7db0a3614995becc25d9a8dce787", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c297fe21-cd0b-4226-813b-a65d2358d034", "external-id": "nsx-vlan-transportzone-98", "segmentation_id": 98, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5d578b0-1f", "ovs_interfaceid": "d5d578b0-1f4a-4a46-8fac-1e625d48c2bc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.781146] env[63531]: DEBUG oslo_vmware.api [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Task: {'id': task-1117984, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.818526] env[63531]: DEBUG nova.compute.manager [req-75dc2d68-054f-4026-8999-b01e3b130f24 req-669735a5-d030-4251-8d49-491eae539f88 service nova] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Received event network-vif-plugged-d5d578b0-1f4a-4a46-8fac-1e625d48c2bc {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 860.818526] env[63531]: DEBUG oslo_concurrency.lockutils [req-75dc2d68-054f-4026-8999-b01e3b130f24 req-669735a5-d030-4251-8d49-491eae539f88 service nova] Acquiring lock "879924d4-b465-4102-a0e3-c7b2be7ef08b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.818526] env[63531]: DEBUG oslo_concurrency.lockutils [req-75dc2d68-054f-4026-8999-b01e3b130f24 req-669735a5-d030-4251-8d49-491eae539f88 service nova] Lock "879924d4-b465-4102-a0e3-c7b2be7ef08b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.818526] env[63531]: DEBUG oslo_concurrency.lockutils [req-75dc2d68-054f-4026-8999-b01e3b130f24 req-669735a5-d030-4251-8d49-491eae539f88 service nova] Lock "879924d4-b465-4102-a0e3-c7b2be7ef08b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.819612] env[63531]: DEBUG nova.compute.manager [req-75dc2d68-054f-4026-8999-b01e3b130f24 req-669735a5-d030-4251-8d49-491eae539f88 service nova] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] No waiting events found dispatching network-vif-plugged-d5d578b0-1f4a-4a46-8fac-1e625d48c2bc {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 860.820820] env[63531]: WARNING nova.compute.manager [req-75dc2d68-054f-4026-8999-b01e3b130f24 req-669735a5-d030-4251-8d49-491eae539f88 service nova] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Received unexpected event network-vif-plugged-d5d578b0-1f4a-4a46-8fac-1e625d48c2bc for instance with vm_state building and task_state spawning. [ 860.820820] env[63531]: DEBUG nova.compute.manager [req-75dc2d68-054f-4026-8999-b01e3b130f24 req-669735a5-d030-4251-8d49-491eae539f88 service nova] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Received event network-changed-d5d578b0-1f4a-4a46-8fac-1e625d48c2bc {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 860.820820] env[63531]: DEBUG nova.compute.manager [req-75dc2d68-054f-4026-8999-b01e3b130f24 req-669735a5-d030-4251-8d49-491eae539f88 service nova] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Refreshing instance network info cache due to event network-changed-d5d578b0-1f4a-4a46-8fac-1e625d48c2bc. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 860.820820] env[63531]: DEBUG oslo_concurrency.lockutils [req-75dc2d68-054f-4026-8999-b01e3b130f24 req-669735a5-d030-4251-8d49-491eae539f88 service nova] Acquiring lock "refresh_cache-879924d4-b465-4102-a0e3-c7b2be7ef08b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.837044] env[63531]: DEBUG oslo_vmware.api [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1117977, 'name': MoveVirtualDisk_Task} progress is 71%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.912553] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ead7921a-cb8c-490f-a3ab-414729b031bc tempest-AttachInterfacesUnderV243Test-838337734 tempest-AttachInterfacesUnderV243Test-838337734-project-member] Lock "0197e03d-6c36-4e73-9472-8e3319eb89ce" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.361s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.932660] env[63531]: DEBUG oslo_vmware.api [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52543ccc-6756-ee12-e397-287ff06a953a, 'name': SearchDatastore_Task, 'duration_secs': 0.075506} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.935902] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fb84ccef-7caf-46d5-91d9-128258ea85b8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.944212] env[63531]: DEBUG oslo_vmware.api [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 860.944212] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52009cd8-feb3-691d-8874-d2f7be8a8e3c" [ 860.944212] env[63531]: _type = "Task" [ 860.944212] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.957191] env[63531]: DEBUG oslo_vmware.api [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52009cd8-feb3-691d-8874-d2f7be8a8e3c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.012024] env[63531]: DEBUG oslo_vmware.api [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52a57765-4084-2ee7-21b4-fe2bc9b654ef, 'name': SearchDatastore_Task, 'duration_secs': 0.016823} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.013145] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-05ffa8c9-8ecb-45dd-a6f0-2458be6b320d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.020177] env[63531]: DEBUG oslo_vmware.api [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 861.020177] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52d031a2-3750-acc2-7e49-00c7be17b101" [ 861.020177] env[63531]: _type = "Task" [ 861.020177] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.031248] env[63531]: DEBUG oslo_vmware.api [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52d031a2-3750-acc2-7e49-00c7be17b101, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.176032] env[63531]: DEBUG nova.network.neutron [None req-1b340bf7-08de-4d21-a861-c6e587976b10 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 861.230432] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Releasing lock "refresh_cache-879924d4-b465-4102-a0e3-c7b2be7ef08b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.230432] env[63531]: DEBUG nova.compute.manager [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Instance network_info: |[{"id": "d5d578b0-1f4a-4a46-8fac-1e625d48c2bc", "address": "fa:16:3e:ca:1a:aa", "network": {"id": "913045d9-e90b-4366-987c-7706b73a6147", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1696510389-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f71c7db0a3614995becc25d9a8dce787", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c297fe21-cd0b-4226-813b-a65d2358d034", "external-id": "nsx-vlan-transportzone-98", "segmentation_id": 98, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5d578b0-1f", "ovs_interfaceid": "d5d578b0-1f4a-4a46-8fac-1e625d48c2bc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 861.230432] env[63531]: DEBUG oslo_concurrency.lockutils [req-75dc2d68-054f-4026-8999-b01e3b130f24 req-669735a5-d030-4251-8d49-491eae539f88 service nova] Acquired lock "refresh_cache-879924d4-b465-4102-a0e3-c7b2be7ef08b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.230432] env[63531]: DEBUG nova.network.neutron [req-75dc2d68-054f-4026-8999-b01e3b130f24 req-669735a5-d030-4251-8d49-491eae539f88 service nova] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Refreshing network info cache for port d5d578b0-1f4a-4a46-8fac-1e625d48c2bc {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 861.231042] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ca:1a:aa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c297fe21-cd0b-4226-813b-a65d2358d034', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd5d578b0-1f4a-4a46-8fac-1e625d48c2bc', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 861.239944] env[63531]: DEBUG oslo.service.loopingcall [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 861.240460] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 861.241786] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-28a77d4c-d5ac-4b7d-971f-757edff60450 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.267174] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 861.267174] env[63531]: value = "task-1117985" [ 861.267174] env[63531]: _type = "Task" [ 861.267174] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.280319] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117985, 'name': CreateVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.284012] env[63531]: DEBUG oslo_vmware.api [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Task: {'id': task-1117984, 'name': PowerOnVM_Task, 'duration_secs': 0.70573} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.284521] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 861.284776] env[63531]: INFO nova.compute.manager [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Took 12.37 seconds to spawn the instance on the hypervisor. [ 861.284971] env[63531]: DEBUG nova.compute.manager [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 861.285901] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3156876-4b6e-47c8-8a4c-097e941cf8f3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.301102] env[63531]: DEBUG nova.network.neutron [None req-1b340bf7-08de-4d21-a861-c6e587976b10 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.338392] env[63531]: DEBUG oslo_vmware.api [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1117977, 'name': MoveVirtualDisk_Task} progress is 94%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.460374] env[63531]: DEBUG oslo_vmware.api [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52009cd8-feb3-691d-8874-d2f7be8a8e3c, 'name': SearchDatastore_Task, 'duration_secs': 0.076414} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.464185] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.464578] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] 2b0e4285-310a-439c-bc50-000a7c5ef7f9/2b0e4285-310a-439c-bc50-000a7c5ef7f9.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 861.465305] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e7494a8a-e809-480a-a28f-78155f035a7f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.473033] env[63531]: DEBUG oslo_vmware.api [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 861.473033] env[63531]: value = "task-1117986" [ 861.473033] env[63531]: _type = "Task" [ 861.473033] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.485706] env[63531]: DEBUG oslo_vmware.api [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1117986, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.530793] env[63531]: DEBUG oslo_vmware.api [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52d031a2-3750-acc2-7e49-00c7be17b101, 'name': SearchDatastore_Task, 'duration_secs': 0.01736} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.533829] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.534140] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df/4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 861.534667] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-68e4bb6e-cb00-4ada-8e0a-46147ea4cbdf {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.541561] env[63531]: DEBUG oslo_vmware.api [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 861.541561] env[63531]: value = "task-1117987" [ 861.541561] env[63531]: _type = "Task" [ 861.541561] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.553372] env[63531]: DEBUG oslo_vmware.api [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1117987, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.779894] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1117985, 'name': CreateVM_Task, 'duration_secs': 0.460004} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.779894] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 861.780392] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 861.780581] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.780884] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 861.781163] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eb72070f-3066-4bce-88d3-a9c19034d73c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.788896] env[63531]: DEBUG oslo_vmware.api [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 861.788896] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52c044f4-c4fb-20d5-e991-877c9d58fa72" [ 861.788896] env[63531]: _type = "Task" [ 861.788896] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.799078] env[63531]: DEBUG oslo_vmware.api [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52c044f4-c4fb-20d5-e991-877c9d58fa72, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.801290] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb58f312-55ee-4c46-8ae2-ee9f95fd1e0e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.808990] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1b340bf7-08de-4d21-a861-c6e587976b10 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Releasing lock "refresh_cache-09feb12a-da9f-4bb7-959d-84c5c7c58e99" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.809597] env[63531]: DEBUG nova.compute.manager [None req-1b340bf7-08de-4d21-a861-c6e587976b10 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 861.809767] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-1b340bf7-08de-4d21-a861-c6e587976b10 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 861.812931] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9beb2ae8-0085-4489-8fa0-5b44ae0206ff {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.817542] env[63531]: INFO nova.compute.manager [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Took 45.06 seconds to build instance. [ 861.826376] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e2df0e2-1fb0-4a6a-b12d-b755f5b3287a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.830581] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b340bf7-08de-4d21-a861-c6e587976b10 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 861.837204] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6f9b0954-7ada-4993-bca7-cedca32d44bb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.847202] env[63531]: DEBUG oslo_vmware.api [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1117977, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.789231} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.877367] env[63531]: INFO nova.virt.vmwareapi.ds_util [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_c5b7f03a-1101-4d08-b9a5-cb3358d88051/OSTACK_IMG_c5b7f03a-1101-4d08-b9a5-cb3358d88051.vmdk to [datastore2] devstack-image-cache_base/8197d52e-21f1-413d-8ef3-59307facd73d/8197d52e-21f1-413d-8ef3-59307facd73d.vmdk. [ 861.877720] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: db83adf4-6183-4330-b260-77d1f5daf899] Cleaning up location [datastore2] OSTACK_IMG_c5b7f03a-1101-4d08-b9a5-cb3358d88051 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 861.878255] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_c5b7f03a-1101-4d08-b9a5-cb3358d88051 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 861.878403] env[63531]: DEBUG oslo_vmware.api [None req-1b340bf7-08de-4d21-a861-c6e587976b10 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Waiting for the task: (returnval){ [ 861.878403] env[63531]: value = "task-1117988" [ 861.878403] env[63531]: _type = "Task" [ 861.878403] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.882268] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6f477cb5-23fc-4065-873a-b6d31092ce30 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.885118] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad448605-ae66-443b-905b-b32638190bb8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.901113] env[63531]: DEBUG oslo_vmware.api [None req-1b340bf7-08de-4d21-a861-c6e587976b10 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': task-1117988, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.903981] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9d22970-8b70-4017-b91e-295f31287c0b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.908817] env[63531]: DEBUG oslo_vmware.api [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Waiting for the task: (returnval){ [ 861.908817] env[63531]: value = "task-1117989" [ 861.908817] env[63531]: _type = "Task" [ 861.908817] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.922126] env[63531]: DEBUG nova.compute.provider_tree [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 861.933392] env[63531]: DEBUG oslo_vmware.api [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1117989, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.984138] env[63531]: DEBUG oslo_vmware.api [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1117986, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.053375] env[63531]: DEBUG oslo_vmware.api [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1117987, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.063717] env[63531]: DEBUG nova.network.neutron [req-75dc2d68-054f-4026-8999-b01e3b130f24 req-669735a5-d030-4251-8d49-491eae539f88 service nova] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Updated VIF entry in instance network info cache for port d5d578b0-1f4a-4a46-8fac-1e625d48c2bc. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 862.064192] env[63531]: DEBUG nova.network.neutron [req-75dc2d68-054f-4026-8999-b01e3b130f24 req-669735a5-d030-4251-8d49-491eae539f88 service nova] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Updating instance_info_cache with network_info: [{"id": "d5d578b0-1f4a-4a46-8fac-1e625d48c2bc", "address": "fa:16:3e:ca:1a:aa", "network": {"id": "913045d9-e90b-4366-987c-7706b73a6147", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1696510389-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f71c7db0a3614995becc25d9a8dce787", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c297fe21-cd0b-4226-813b-a65d2358d034", "external-id": "nsx-vlan-transportzone-98", "segmentation_id": 98, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5d578b0-1f", "ovs_interfaceid": "d5d578b0-1f4a-4a46-8fac-1e625d48c2bc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 862.301415] env[63531]: DEBUG oslo_vmware.api [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52c044f4-c4fb-20d5-e991-877c9d58fa72, 'name': SearchDatastore_Task, 'duration_secs': 0.017559} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.301748] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.302087] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 862.302388] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.302547] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.302801] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 862.303155] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b480bfff-411e-47c9-827f-a58c122b5aea {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.315287] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 862.315512] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 862.316512] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8ff9315f-3285-4788-924e-d1dd7f4a1558 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.319763] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8d7ba8a6-ab0a-4875-9e35-514099661031 tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Lock "92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 52.820s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.323671] env[63531]: DEBUG oslo_vmware.api [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 862.323671] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]527195f7-f956-9e4d-bf3d-62b7d9a5935e" [ 862.323671] env[63531]: _type = "Task" [ 862.323671] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.333613] env[63531]: DEBUG oslo_vmware.api [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]527195f7-f956-9e4d-bf3d-62b7d9a5935e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.393530] env[63531]: DEBUG oslo_vmware.api [None req-1b340bf7-08de-4d21-a861-c6e587976b10 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': task-1117988, 'name': PowerOffVM_Task, 'duration_secs': 0.214862} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.393843] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b340bf7-08de-4d21-a861-c6e587976b10 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 862.394021] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-1b340bf7-08de-4d21-a861-c6e587976b10 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 862.394317] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-56cff57e-d5ef-47b5-9e0f-97bd31b46586 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.421254] env[63531]: DEBUG oslo_vmware.api [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1117989, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.232842} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.422238] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 862.422471] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8197d52e-21f1-413d-8ef3-59307facd73d/8197d52e-21f1-413d-8ef3-59307facd73d.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.422782] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8197d52e-21f1-413d-8ef3-59307facd73d/8197d52e-21f1-413d-8ef3-59307facd73d.vmdk to [datastore2] db83adf4-6183-4330-b260-77d1f5daf899/db83adf4-6183-4330-b260-77d1f5daf899.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 862.423142] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-1b340bf7-08de-4d21-a861-c6e587976b10 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 862.423377] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-1b340bf7-08de-4d21-a861-c6e587976b10 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Deleting contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 862.423599] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-1b340bf7-08de-4d21-a861-c6e587976b10 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Deleting the datastore file [datastore2] 09feb12a-da9f-4bb7-959d-84c5c7c58e99 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 862.423900] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ce0b4c62-6b99-44ce-9336-2c45dc588173 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.426230] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c7c010d0-0915-40f8-9044-76f49c458598 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.428985] env[63531]: DEBUG nova.scheduler.client.report [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 862.438766] env[63531]: DEBUG oslo_vmware.api [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Waiting for the task: (returnval){ [ 862.438766] env[63531]: value = "task-1117991" [ 862.438766] env[63531]: _type = "Task" [ 862.438766] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.439104] env[63531]: DEBUG oslo_vmware.api [None req-1b340bf7-08de-4d21-a861-c6e587976b10 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Waiting for the task: (returnval){ [ 862.439104] env[63531]: value = "task-1117992" [ 862.439104] env[63531]: _type = "Task" [ 862.439104] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.451789] env[63531]: DEBUG oslo_vmware.api [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1117991, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.455275] env[63531]: DEBUG oslo_vmware.api [None req-1b340bf7-08de-4d21-a861-c6e587976b10 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': task-1117992, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.483752] env[63531]: DEBUG oslo_vmware.api [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1117986, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.719318} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.484242] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] 2b0e4285-310a-439c-bc50-000a7c5ef7f9/2b0e4285-310a-439c-bc50-000a7c5ef7f9.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 862.484340] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 862.484635] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-de2917b0-5f6b-42ad-b8e5-32e8b992d8b5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.491153] env[63531]: DEBUG oslo_vmware.api [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 862.491153] env[63531]: value = "task-1117993" [ 862.491153] env[63531]: _type = "Task" [ 862.491153] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.499704] env[63531]: DEBUG oslo_vmware.api [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1117993, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.552270] env[63531]: DEBUG oslo_vmware.api [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1117987, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.802737} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.552594] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df/4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 862.552849] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 862.553131] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a03f0fb8-2089-4eb1-b6e3-5db73ee29804 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.559993] env[63531]: DEBUG oslo_vmware.api [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 862.559993] env[63531]: value = "task-1117994" [ 862.559993] env[63531]: _type = "Task" [ 862.559993] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.568294] env[63531]: DEBUG oslo_concurrency.lockutils [req-75dc2d68-054f-4026-8999-b01e3b130f24 req-669735a5-d030-4251-8d49-491eae539f88 service nova] Releasing lock "refresh_cache-879924d4-b465-4102-a0e3-c7b2be7ef08b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.568855] env[63531]: DEBUG oslo_vmware.api [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1117994, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.837103] env[63531]: DEBUG oslo_vmware.api [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]527195f7-f956-9e4d-bf3d-62b7d9a5935e, 'name': SearchDatastore_Task, 'duration_secs': 0.010241} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.838031] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3525b895-d101-4820-9bf5-8858d99b24e9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.846813] env[63531]: DEBUG oslo_vmware.api [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 862.846813] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52c93acc-747c-9f8b-c36a-3be8633d8be9" [ 862.846813] env[63531]: _type = "Task" [ 862.846813] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.859175] env[63531]: DEBUG oslo_vmware.api [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52c93acc-747c-9f8b-c36a-3be8633d8be9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.935056] env[63531]: DEBUG oslo_concurrency.lockutils [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.575s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.935578] env[63531]: DEBUG nova.compute.manager [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 862.938504] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f49106e4-b7c8-427f-86d8-e698c2d39d08 tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.646s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.938745] env[63531]: DEBUG nova.objects.instance [None req-f49106e4-b7c8-427f-86d8-e698c2d39d08 tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Lazy-loading 'resources' on Instance uuid 43c29443-1210-4ee2-95c7-1257de308287 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 862.954203] env[63531]: DEBUG oslo_vmware.api [None req-1b340bf7-08de-4d21-a861-c6e587976b10 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': task-1117992, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.161956} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.960279] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-1b340bf7-08de-4d21-a861-c6e587976b10 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 862.960598] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-1b340bf7-08de-4d21-a861-c6e587976b10 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Deleted contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 862.960806] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-1b340bf7-08de-4d21-a861-c6e587976b10 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 862.961160] env[63531]: INFO nova.compute.manager [None req-1b340bf7-08de-4d21-a861-c6e587976b10 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Took 1.15 seconds to destroy the instance on the hypervisor. [ 862.961630] env[63531]: DEBUG oslo.service.loopingcall [None req-1b340bf7-08de-4d21-a861-c6e587976b10 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 862.962412] env[63531]: DEBUG oslo_vmware.api [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1117991, 'name': CopyVirtualDisk_Task} progress is 18%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.962755] env[63531]: DEBUG nova.compute.manager [-] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 862.962860] env[63531]: DEBUG nova.network.neutron [-] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 862.991917] env[63531]: DEBUG nova.network.neutron [-] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 863.008919] env[63531]: DEBUG oslo_vmware.api [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1117993, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.088468} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.009307] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 863.010422] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08370ea4-2b8d-4eb1-96f7-bacd5d44080b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.038433] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Reconfiguring VM instance instance-00000043 to attach disk [datastore2] 2b0e4285-310a-439c-bc50-000a7c5ef7f9/2b0e4285-310a-439c-bc50-000a7c5ef7f9.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 863.039295] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2c9d60b6-f7b0-4763-8487-a3c7832076da {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.067971] env[63531]: DEBUG oslo_vmware.api [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 863.067971] env[63531]: value = "task-1117995" [ 863.067971] env[63531]: _type = "Task" [ 863.067971] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.073841] env[63531]: DEBUG oslo_vmware.api [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1117994, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069731} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.074673] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 863.075646] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-890801ce-8329-49c8-953c-b14691e3b740 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.083706] env[63531]: DEBUG oslo_vmware.api [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1117995, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.107754] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Reconfiguring VM instance instance-00000046 to attach disk [datastore1] 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df/4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 863.108162] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a11dba89-145c-4cd9-a15f-f39dcc9a9630 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.128290] env[63531]: DEBUG oslo_vmware.api [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 863.128290] env[63531]: value = "task-1117996" [ 863.128290] env[63531]: _type = "Task" [ 863.128290] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.140460] env[63531]: DEBUG oslo_vmware.api [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1117996, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.366265] env[63531]: DEBUG oslo_vmware.api [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52c93acc-747c-9f8b-c36a-3be8633d8be9, 'name': SearchDatastore_Task, 'duration_secs': 0.071633} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.366655] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.366962] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] 879924d4-b465-4102-a0e3-c7b2be7ef08b/879924d4-b465-4102-a0e3-c7b2be7ef08b.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 863.367390] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7ddb0549-7efc-4ad4-bc54-d819dc647b5b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.378511] env[63531]: DEBUG oslo_vmware.api [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 863.378511] env[63531]: value = "task-1117997" [ 863.378511] env[63531]: _type = "Task" [ 863.378511] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.388468] env[63531]: DEBUG oslo_vmware.api [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1117997, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.442230] env[63531]: DEBUG nova.compute.utils [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 863.444455] env[63531]: DEBUG nova.compute.manager [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 863.444832] env[63531]: DEBUG nova.network.neutron [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 863.464956] env[63531]: DEBUG oslo_vmware.api [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1117991, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.499478] env[63531]: DEBUG nova.network.neutron [-] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 863.528183] env[63531]: DEBUG nova.policy [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0a501b2c64c04dcda3596c8a1bd3fa0d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'abdae960d43b47bc95ab41f265d2c85b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 863.594198] env[63531]: DEBUG oslo_vmware.api [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1117995, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.644268] env[63531]: DEBUG oslo_vmware.api [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1117996, 'name': ReconfigVM_Task, 'duration_secs': 0.488373} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.644763] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Reconfigured VM instance instance-00000046 to attach disk [datastore1] 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df/4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 863.645952] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-78b563ce-0b83-444c-9ac6-b4d84904f42e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.653189] env[63531]: DEBUG oslo_vmware.api [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 863.653189] env[63531]: value = "task-1117998" [ 863.653189] env[63531]: _type = "Task" [ 863.653189] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.669168] env[63531]: DEBUG oslo_vmware.api [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1117998, 'name': Rename_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.896302] env[63531]: DEBUG oslo_vmware.api [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1117997, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.903300] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e23a84c5-7c5f-4747-baff-f99f8bcc41e9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.913136] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-723c96e8-fb91-4024-971f-61ad10dc834d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.955563] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-821d54c8-f5b6-465d-835e-33469cb6a52c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.961088] env[63531]: DEBUG nova.compute.manager [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 863.972370] env[63531]: DEBUG oslo_vmware.api [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1117991, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.978667] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5966a99-1b19-4652-8eb4-c9f65bf7b571 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.999409] env[63531]: DEBUG nova.compute.provider_tree [None req-f49106e4-b7c8-427f-86d8-e698c2d39d08 tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 864.005017] env[63531]: INFO nova.compute.manager [-] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Took 1.04 seconds to deallocate network for instance. [ 864.079560] env[63531]: DEBUG oslo_vmware.api [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1117995, 'name': ReconfigVM_Task, 'duration_secs': 0.537245} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.080672] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Reconfigured VM instance instance-00000043 to attach disk [datastore2] 2b0e4285-310a-439c-bc50-000a7c5ef7f9/2b0e4285-310a-439c-bc50-000a7c5ef7f9.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 864.081626] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0bfc1fea-8b01-4a39-a5f0-259eb07e3b53 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.089911] env[63531]: DEBUG oslo_vmware.api [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 864.089911] env[63531]: value = "task-1117999" [ 864.089911] env[63531]: _type = "Task" [ 864.089911] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.094018] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f2e31732-dbaf-4d48-baef-afe717b5167a tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Acquiring lock "92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.094641] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f2e31732-dbaf-4d48-baef-afe717b5167a tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Lock "92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.094931] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f2e31732-dbaf-4d48-baef-afe717b5167a tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Acquiring lock "92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.095681] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f2e31732-dbaf-4d48-baef-afe717b5167a tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Lock "92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.095681] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f2e31732-dbaf-4d48-baef-afe717b5167a tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Lock "92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.101352] env[63531]: INFO nova.compute.manager [None req-f2e31732-dbaf-4d48-baef-afe717b5167a tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Terminating instance [ 864.107765] env[63531]: DEBUG oslo_vmware.api [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1117999, 'name': Rename_Task} progress is 10%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.108630] env[63531]: DEBUG nova.compute.manager [None req-f2e31732-dbaf-4d48-baef-afe717b5167a tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 864.108827] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f2e31732-dbaf-4d48-baef-afe717b5167a tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 864.110314] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66d9e4bb-719a-4911-87cd-c2ed62fe6247 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.118894] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2e31732-dbaf-4d48-baef-afe717b5167a tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 864.119350] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-106a7fbb-c8eb-4472-a359-ab3d86fccd5f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.129336] env[63531]: DEBUG oslo_vmware.api [None req-f2e31732-dbaf-4d48-baef-afe717b5167a tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Waiting for the task: (returnval){ [ 864.129336] env[63531]: value = "task-1118000" [ 864.129336] env[63531]: _type = "Task" [ 864.129336] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.142142] env[63531]: DEBUG oslo_vmware.api [None req-f2e31732-dbaf-4d48-baef-afe717b5167a tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Task: {'id': task-1118000, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.152334] env[63531]: DEBUG nova.network.neutron [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Successfully created port: a806c2a6-5a84-4114-aedb-8bb2da53a147 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 864.171565] env[63531]: DEBUG oslo_vmware.api [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1117998, 'name': Rename_Task, 'duration_secs': 0.177212} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.171940] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 864.172440] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6df67a88-a5fa-45a8-8614-bacd7072e792 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.182525] env[63531]: DEBUG oslo_vmware.api [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 864.182525] env[63531]: value = "task-1118001" [ 864.182525] env[63531]: _type = "Task" [ 864.182525] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.197653] env[63531]: DEBUG oslo_vmware.api [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118001, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.392466] env[63531]: DEBUG oslo_vmware.api [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1117997, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.464752] env[63531]: DEBUG oslo_vmware.api [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1117991, 'name': CopyVirtualDisk_Task} progress is 85%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.502256] env[63531]: DEBUG nova.scheduler.client.report [None req-f49106e4-b7c8-427f-86d8-e698c2d39d08 tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 864.512472] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1b340bf7-08de-4d21-a861-c6e587976b10 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.606048] env[63531]: DEBUG oslo_vmware.api [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1117999, 'name': Rename_Task, 'duration_secs': 0.227171} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.606048] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 864.606048] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e1aca8a5-5266-4e38-a568-d1f032926bae {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.619464] env[63531]: DEBUG oslo_vmware.api [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 864.619464] env[63531]: value = "task-1118002" [ 864.619464] env[63531]: _type = "Task" [ 864.619464] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.635029] env[63531]: DEBUG oslo_vmware.api [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118002, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.650248] env[63531]: DEBUG oslo_vmware.api [None req-f2e31732-dbaf-4d48-baef-afe717b5167a tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Task: {'id': task-1118000, 'name': PowerOffVM_Task, 'duration_secs': 0.247809} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.650764] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2e31732-dbaf-4d48-baef-afe717b5167a tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 864.650979] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f2e31732-dbaf-4d48-baef-afe717b5167a tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 864.651700] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e0f141f2-58bd-4238-b7a6-ce5a477229e4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.696744] env[63531]: DEBUG oslo_vmware.api [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118001, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.774271] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f2e31732-dbaf-4d48-baef-afe717b5167a tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 864.774578] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f2e31732-dbaf-4d48-baef-afe717b5167a tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Deleting contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 864.774794] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2e31732-dbaf-4d48-baef-afe717b5167a tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Deleting the datastore file [datastore2] 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 864.775280] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b791480e-6909-40f8-bfbb-85cd7c69a02c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.784756] env[63531]: DEBUG oslo_vmware.api [None req-f2e31732-dbaf-4d48-baef-afe717b5167a tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Waiting for the task: (returnval){ [ 864.784756] env[63531]: value = "task-1118004" [ 864.784756] env[63531]: _type = "Task" [ 864.784756] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.796115] env[63531]: DEBUG oslo_vmware.api [None req-f2e31732-dbaf-4d48-baef-afe717b5167a tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Task: {'id': task-1118004, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.900550] env[63531]: DEBUG oslo_vmware.api [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1117997, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.966336] env[63531]: DEBUG oslo_vmware.api [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1117991, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.365653} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.967729] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8197d52e-21f1-413d-8ef3-59307facd73d/8197d52e-21f1-413d-8ef3-59307facd73d.vmdk to [datastore2] db83adf4-6183-4330-b260-77d1f5daf899/db83adf4-6183-4330-b260-77d1f5daf899.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 864.967964] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af1b9b10-5eeb-412d-ba5b-576e42a5c3c2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.976261] env[63531]: DEBUG nova.compute.manager [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 865.001158] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: db83adf4-6183-4330-b260-77d1f5daf899] Reconfiguring VM instance instance-00000045 to attach disk [datastore2] db83adf4-6183-4330-b260-77d1f5daf899/db83adf4-6183-4330-b260-77d1f5daf899.vmdk or device None with type streamOptimized {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 865.002939] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c94056ea-5437-46bd-82e9-5017fb00a9b4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.023455] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f49106e4-b7c8-427f-86d8-e698c2d39d08 tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.085s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.029087] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a81f1a0b-6718-4880-9962-0449389e3290 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.892s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.029087] env[63531]: DEBUG nova.objects.instance [None req-a81f1a0b-6718-4880-9962-0449389e3290 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Lazy-loading 'resources' on Instance uuid e0074e2e-e8bd-4bcb-9ad9-271a10e079e4 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 865.038579] env[63531]: DEBUG oslo_vmware.api [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Waiting for the task: (returnval){ [ 865.038579] env[63531]: value = "task-1118005" [ 865.038579] env[63531]: _type = "Task" [ 865.038579] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.054370] env[63531]: DEBUG oslo_vmware.api [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1118005, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.064208] env[63531]: INFO nova.scheduler.client.report [None req-f49106e4-b7c8-427f-86d8-e698c2d39d08 tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Deleted allocations for instance 43c29443-1210-4ee2-95c7-1257de308287 [ 865.136643] env[63531]: DEBUG oslo_vmware.api [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118002, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.201808] env[63531]: DEBUG oslo_vmware.api [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118001, 'name': PowerOnVM_Task, 'duration_secs': 0.710609} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.201808] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 865.201808] env[63531]: INFO nova.compute.manager [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Took 9.03 seconds to spawn the instance on the hypervisor. [ 865.201808] env[63531]: DEBUG nova.compute.manager [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 865.202685] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcc1b580-27ae-4e7a-b6d6-e2616c7963fa {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.308047] env[63531]: DEBUG oslo_vmware.api [None req-f2e31732-dbaf-4d48-baef-afe717b5167a tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Task: {'id': task-1118004, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.499591} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.308047] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2e31732-dbaf-4d48-baef-afe717b5167a tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 865.308047] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f2e31732-dbaf-4d48-baef-afe717b5167a tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Deleted contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 865.308047] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f2e31732-dbaf-4d48-baef-afe717b5167a tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 865.308047] env[63531]: INFO nova.compute.manager [None req-f2e31732-dbaf-4d48-baef-afe717b5167a tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Took 1.20 seconds to destroy the instance on the hypervisor. [ 865.308047] env[63531]: DEBUG oslo.service.loopingcall [None req-f2e31732-dbaf-4d48-baef-afe717b5167a tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 865.312190] env[63531]: DEBUG nova.compute.manager [-] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 865.315323] env[63531]: DEBUG nova.network.neutron [-] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 865.397505] env[63531]: DEBUG oslo_vmware.api [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1117997, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.771493} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.397740] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] 879924d4-b465-4102-a0e3-c7b2be7ef08b/879924d4-b465-4102-a0e3-c7b2be7ef08b.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 865.397993] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 865.398282] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-160e7113-3305-4a9a-bbb5-e2a6b8b41e28 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.409881] env[63531]: DEBUG oslo_vmware.api [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 865.409881] env[63531]: value = "task-1118006" [ 865.409881] env[63531]: _type = "Task" [ 865.409881] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.424304] env[63531]: DEBUG oslo_vmware.api [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118006, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.550013] env[63531]: DEBUG oslo_vmware.api [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1118005, 'name': ReconfigVM_Task, 'duration_secs': 0.48781} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.550375] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: db83adf4-6183-4330-b260-77d1f5daf899] Reconfigured VM instance instance-00000045 to attach disk [datastore2] db83adf4-6183-4330-b260-77d1f5daf899/db83adf4-6183-4330-b260-77d1f5daf899.vmdk or device None with type streamOptimized {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 865.551017] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ba77aaae-cce9-49ca-8ccc-d0ec55a69229 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.565149] env[63531]: DEBUG oslo_vmware.api [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Waiting for the task: (returnval){ [ 865.565149] env[63531]: value = "task-1118007" [ 865.565149] env[63531]: _type = "Task" [ 865.565149] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.579652] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f49106e4-b7c8-427f-86d8-e698c2d39d08 tempest-ServerAddressesNegativeTestJSON-838134913 tempest-ServerAddressesNegativeTestJSON-838134913-project-member] Lock "43c29443-1210-4ee2-95c7-1257de308287" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.383s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.590213] env[63531]: DEBUG oslo_vmware.api [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1118007, 'name': Rename_Task} progress is 10%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.643599] env[63531]: DEBUG oslo_vmware.api [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118002, 'name': PowerOnVM_Task, 'duration_secs': 0.728668} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.643985] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 865.644191] env[63531]: DEBUG nova.compute.manager [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 865.645104] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d60156d1-911b-465e-adcb-1cced96f8719 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.728321] env[63531]: INFO nova.compute.manager [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Took 38.88 seconds to build instance. [ 865.733654] env[63531]: DEBUG nova.virt.hardware [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 865.733958] env[63531]: DEBUG nova.virt.hardware [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 865.734133] env[63531]: DEBUG nova.virt.hardware [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 865.734352] env[63531]: DEBUG nova.virt.hardware [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 865.734465] env[63531]: DEBUG nova.virt.hardware [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 865.734614] env[63531]: DEBUG nova.virt.hardware [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 865.734867] env[63531]: DEBUG nova.virt.hardware [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 865.735571] env[63531]: DEBUG nova.virt.hardware [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 865.735823] env[63531]: DEBUG nova.virt.hardware [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 865.736031] env[63531]: DEBUG nova.virt.hardware [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 865.736220] env[63531]: DEBUG nova.virt.hardware [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 865.738627] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fa80ff3-a1bc-479b-9688-1539e48cc015 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.753686] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5f35cda-bfda-4e9a-a8bb-5d5e037fbdb0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.780223] env[63531]: DEBUG oslo_vmware.rw_handles [None req-dba9d971-47b5-433d-b369-13c4616060ba tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528e465a-6ee9-3b42-7d15-93f9c0e4c59a/disk-0.vmdk. {{(pid=63531) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 865.781568] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4968ecbc-e1f4-4089-bf61-d3f01e94ed72 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.793292] env[63531]: DEBUG oslo_vmware.rw_handles [None req-dba9d971-47b5-433d-b369-13c4616060ba tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528e465a-6ee9-3b42-7d15-93f9c0e4c59a/disk-0.vmdk is in state: ready. {{(pid=63531) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 865.793477] env[63531]: ERROR oslo_vmware.rw_handles [None req-dba9d971-47b5-433d-b369-13c4616060ba tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528e465a-6ee9-3b42-7d15-93f9c0e4c59a/disk-0.vmdk due to incomplete transfer. [ 865.793990] env[63531]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-7b5694d4-c725-4da1-b9bf-57e952f598b4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.818100] env[63531]: DEBUG oslo_vmware.rw_handles [None req-dba9d971-47b5-433d-b369-13c4616060ba tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528e465a-6ee9-3b42-7d15-93f9c0e4c59a/disk-0.vmdk. {{(pid=63531) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 865.818100] env[63531]: DEBUG nova.virt.vmwareapi.images [None req-dba9d971-47b5-433d-b369-13c4616060ba tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Uploaded image 2d8829b0-25e0-44ee-a350-e1cd7eb3e470 to the Glance image server {{(pid=63531) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 865.818100] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-dba9d971-47b5-433d-b369-13c4616060ba tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Destroying the VM {{(pid=63531) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 865.818371] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-e6bd84a3-0689-46c4-b1c6-d79226e53c03 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.826922] env[63531]: DEBUG oslo_vmware.api [None req-dba9d971-47b5-433d-b369-13c4616060ba tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 865.826922] env[63531]: value = "task-1118008" [ 865.826922] env[63531]: _type = "Task" [ 865.826922] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.844635] env[63531]: DEBUG oslo_vmware.api [None req-dba9d971-47b5-433d-b369-13c4616060ba tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118008, 'name': Destroy_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.924541] env[63531]: DEBUG oslo_vmware.api [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118006, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.136475} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.924944] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 865.925944] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e255fa19-e272-4f66-8217-a997b5bc948d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.954854] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Reconfiguring VM instance instance-00000047 to attach disk [datastore2] 879924d4-b465-4102-a0e3-c7b2be7ef08b/879924d4-b465-4102-a0e3-c7b2be7ef08b.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 865.960456] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4ea02d87-9f6a-46ec-b152-2215a97d6349 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.984035] env[63531]: DEBUG oslo_vmware.api [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 865.984035] env[63531]: value = "task-1118009" [ 865.984035] env[63531]: _type = "Task" [ 865.984035] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.993175] env[63531]: DEBUG oslo_vmware.api [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118009, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.069292] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e453c47-bb5f-4e5f-ba37-c7b84a87a574 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.085974] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4b6a4f0-5093-47e4-b50b-cab0b5b31cfc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.089783] env[63531]: DEBUG oslo_vmware.api [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1118007, 'name': Rename_Task, 'duration_secs': 0.210996} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.090340] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: db83adf4-6183-4330-b260-77d1f5daf899] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 866.091441] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-327ed3b8-dd7f-4790-8227-8bbbbc73fa08 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.903726] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4ddba638-011f-4d48-a17f-3da1fc32fc28 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Lock "4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.272s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.908332] env[63531]: DEBUG nova.network.neutron [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Successfully updated port: a806c2a6-5a84-4114-aedb-8bb2da53a147 {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 866.918537] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f773e4ad-0e85-440f-a52d-a6a5d49748fc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.922308] env[63531]: DEBUG nova.compute.manager [req-8ad84f20-94e7-4286-9b10-e26f10fb815e req-46272195-1621-488a-999c-444ece10e38e service nova] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Received event network-vif-deleted-b9ea99ca-f03c-4164-b8ae-b68245e39a5a {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 866.922497] env[63531]: INFO nova.compute.manager [req-8ad84f20-94e7-4286-9b10-e26f10fb815e req-46272195-1621-488a-999c-444ece10e38e service nova] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Neutron deleted interface b9ea99ca-f03c-4164-b8ae-b68245e39a5a; detaching it from the instance and deleting it from the info cache [ 866.922750] env[63531]: DEBUG nova.network.neutron [req-8ad84f20-94e7-4286-9b10-e26f10fb815e req-46272195-1621-488a-999c-444ece10e38e service nova] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Updating instance_info_cache with network_info: [{"id": "cbb5eb21-bc39-4608-820f-a77b3f5fdedf", "address": "fa:16:3e:d0:eb:72", "network": {"id": "4ae7e465-bcda-457c-a060-64abd40b7152", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-251859226", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.59", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "8bebde0e281d485ebdd70a1a3b613632", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1dc3a0d3-9578-4c45-bc16-13ed5b84f5b4", "external-id": "nsx-vlan-transportzone-817", "segmentation_id": 817, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcbb5eb21-bc", "ovs_interfaceid": "cbb5eb21-bc39-4608-820f-a77b3f5fdedf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 866.925195] env[63531]: DEBUG oslo_vmware.api [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Waiting for the task: (returnval){ [ 866.925195] env[63531]: value = "task-1118010" [ 866.925195] env[63531]: _type = "Task" [ 866.925195] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.925327] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.928548] env[63531]: DEBUG oslo_concurrency.lockutils [None req-545f958f-5555-4fdc-8d01-3ef53b1af173 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Acquiring lock "f46d8a85-6ab7-436a-bc5e-40d0f80974e5" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.928548] env[63531]: DEBUG oslo_concurrency.lockutils [None req-545f958f-5555-4fdc-8d01-3ef53b1af173 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lock "f46d8a85-6ab7-436a-bc5e-40d0f80974e5" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.942205] env[63531]: DEBUG oslo_vmware.api [None req-dba9d971-47b5-433d-b369-13c4616060ba tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118008, 'name': Destroy_Task, 'duration_secs': 0.451604} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.942478] env[63531]: DEBUG oslo_vmware.api [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118009, 'name': ReconfigVM_Task, 'duration_secs': 0.629603} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.944128] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5519f37-3dd3-41d8-a180-700973b97baf {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.952098] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-dba9d971-47b5-433d-b369-13c4616060ba tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Destroyed the VM [ 866.952098] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-dba9d971-47b5-433d-b369-13c4616060ba tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Deleting Snapshot of the VM instance {{(pid=63531) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 866.952311] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Reconfigured VM instance instance-00000047 to attach disk [datastore2] 879924d4-b465-4102-a0e3-c7b2be7ef08b/879924d4-b465-4102-a0e3-c7b2be7ef08b.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 866.959630] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-7d75716e-11dc-4f96-8ebd-646cc565f871 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.959791] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dd130653-1a5f-4031-8004-cc8100101f36 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.967112] env[63531]: DEBUG oslo_vmware.api [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1118010, 'name': PowerOnVM_Task, 'duration_secs': 0.596824} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.967112] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: db83adf4-6183-4330-b260-77d1f5daf899] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 866.967112] env[63531]: INFO nova.compute.manager [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: db83adf4-6183-4330-b260-77d1f5daf899] Took 15.49 seconds to spawn the instance on the hypervisor. [ 866.967112] env[63531]: DEBUG nova.compute.manager [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: db83adf4-6183-4330-b260-77d1f5daf899] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 866.967112] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-398e9ed8-18bc-4431-ad17-bfc674ca11c7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.978138] env[63531]: DEBUG oslo_vmware.api [None req-dba9d971-47b5-433d-b369-13c4616060ba tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 866.978138] env[63531]: value = "task-1118011" [ 866.978138] env[63531]: _type = "Task" [ 866.978138] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.979321] env[63531]: DEBUG nova.compute.provider_tree [None req-a81f1a0b-6718-4880-9962-0449389e3290 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 866.984067] env[63531]: DEBUG oslo_vmware.api [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 866.984067] env[63531]: value = "task-1118012" [ 866.984067] env[63531]: _type = "Task" [ 866.984067] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.005999] env[63531]: DEBUG oslo_vmware.api [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118012, 'name': Rename_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.014545] env[63531]: DEBUG oslo_vmware.api [None req-dba9d971-47b5-433d-b369-13c4616060ba tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118011, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.427037] env[63531]: DEBUG oslo_concurrency.lockutils [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquiring lock "refresh_cache-e95d3273-8216-47cc-95b6-99301366a827" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.427839] env[63531]: DEBUG oslo_concurrency.lockutils [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquired lock "refresh_cache-e95d3273-8216-47cc-95b6-99301366a827" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.428065] env[63531]: DEBUG nova.network.neutron [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 867.429821] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dfb7f38a-77eb-4a20-8d24-0f886b71713e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.440025] env[63531]: DEBUG nova.compute.utils [None req-545f958f-5555-4fdc-8d01-3ef53b1af173 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 867.449951] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7e5084f-c1e5-487c-b632-6412259f4d5a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.486492] env[63531]: DEBUG nova.scheduler.client.report [None req-a81f1a0b-6718-4880-9962-0449389e3290 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 867.515022] env[63531]: DEBUG nova.compute.manager [req-8ad84f20-94e7-4286-9b10-e26f10fb815e req-46272195-1621-488a-999c-444ece10e38e service nova] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Detach interface failed, port_id=b9ea99ca-f03c-4164-b8ae-b68245e39a5a, reason: Instance 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b could not be found. {{(pid=63531) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 867.515022] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a81f1a0b-6718-4880-9962-0449389e3290 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.488s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.527129] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.640s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.530368] env[63531]: INFO nova.compute.claims [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 867.532161] env[63531]: INFO nova.compute.manager [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: db83adf4-6183-4330-b260-77d1f5daf899] Took 44.48 seconds to build instance. [ 867.547035] env[63531]: DEBUG oslo_vmware.api [None req-dba9d971-47b5-433d-b369-13c4616060ba tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118011, 'name': RemoveSnapshot_Task} progress is 56%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.547226] env[63531]: DEBUG oslo_vmware.api [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118012, 'name': Rename_Task, 'duration_secs': 0.223806} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.547738] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 867.548105] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3c501dd1-52ae-4a78-93da-d7d4c37da0c8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.558248] env[63531]: DEBUG oslo_vmware.api [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 867.558248] env[63531]: value = "task-1118013" [ 867.558248] env[63531]: _type = "Task" [ 867.558248] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.573013] env[63531]: DEBUG oslo_vmware.api [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118013, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.574869] env[63531]: INFO nova.scheduler.client.report [None req-a81f1a0b-6718-4880-9962-0449389e3290 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Deleted allocations for instance e0074e2e-e8bd-4bcb-9ad9-271a10e079e4 [ 867.770510] env[63531]: DEBUG nova.network.neutron [-] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 867.951267] env[63531]: DEBUG oslo_concurrency.lockutils [None req-545f958f-5555-4fdc-8d01-3ef53b1af173 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lock "f46d8a85-6ab7-436a-bc5e-40d0f80974e5" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.018s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.978656] env[63531]: DEBUG nova.network.neutron [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 868.014472] env[63531]: DEBUG oslo_vmware.api [None req-dba9d971-47b5-433d-b369-13c4616060ba tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118011, 'name': RemoveSnapshot_Task, 'duration_secs': 0.647523} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.014744] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-dba9d971-47b5-433d-b369-13c4616060ba tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Deleted Snapshot of the VM instance {{(pid=63531) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 868.014988] env[63531]: INFO nova.compute.manager [None req-dba9d971-47b5-433d-b369-13c4616060ba tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Took 16.93 seconds to snapshot the instance on the hypervisor. [ 868.040281] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8a37d2ea-b6d1-470c-91d8-36d49c967654 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Lock "db83adf4-6183-4330-b260-77d1f5daf899" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 57.467s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.070741] env[63531]: DEBUG oslo_vmware.api [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118013, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.084582] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a81f1a0b-6718-4880-9962-0449389e3290 tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Lock "e0074e2e-e8bd-4bcb-9ad9-271a10e079e4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.870s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.135979] env[63531]: DEBUG nova.network.neutron [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Updating instance_info_cache with network_info: [{"id": "a806c2a6-5a84-4114-aedb-8bb2da53a147", "address": "fa:16:3e:1e:bc:64", "network": {"id": "eac8786d-bc67-4916-8483-b53d9a25efb8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1604989690-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "abdae960d43b47bc95ab41f265d2c85b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db068f71-08cc-42d4-8ab6-17134c1585e5", "external-id": "nsx-vlan-transportzone-721", "segmentation_id": 721, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa806c2a6-5a", "ovs_interfaceid": "a806c2a6-5a84-4114-aedb-8bb2da53a147", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 868.275628] env[63531]: INFO nova.compute.manager [-] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Took 2.96 seconds to deallocate network for instance. [ 868.570266] env[63531]: DEBUG nova.compute.manager [None req-dba9d971-47b5-433d-b369-13c4616060ba tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Found 2 images (rotation: 2) {{(pid=63531) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 868.581886] env[63531]: DEBUG oslo_vmware.api [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118013, 'name': PowerOnVM_Task, 'duration_secs': 0.68619} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.583667] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 868.585063] env[63531]: INFO nova.compute.manager [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Took 9.82 seconds to spawn the instance on the hypervisor. [ 868.585717] env[63531]: DEBUG nova.compute.manager [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 868.588071] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d1a9f71-f596-431a-8566-91b3cea577fb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.639766] env[63531]: DEBUG oslo_concurrency.lockutils [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Releasing lock "refresh_cache-e95d3273-8216-47cc-95b6-99301366a827" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 868.640157] env[63531]: DEBUG nova.compute.manager [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Instance network_info: |[{"id": "a806c2a6-5a84-4114-aedb-8bb2da53a147", "address": "fa:16:3e:1e:bc:64", "network": {"id": "eac8786d-bc67-4916-8483-b53d9a25efb8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1604989690-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "abdae960d43b47bc95ab41f265d2c85b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db068f71-08cc-42d4-8ab6-17134c1585e5", "external-id": "nsx-vlan-transportzone-721", "segmentation_id": 721, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa806c2a6-5a", "ovs_interfaceid": "a806c2a6-5a84-4114-aedb-8bb2da53a147", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 868.640599] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1e:bc:64', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'db068f71-08cc-42d4-8ab6-17134c1585e5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a806c2a6-5a84-4114-aedb-8bb2da53a147', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 868.648661] env[63531]: DEBUG oslo.service.loopingcall [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 868.649183] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e95d3273-8216-47cc-95b6-99301366a827] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 868.649183] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3abac47f-7401-42f3-b8ab-dabed0bfca95 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.667173] env[63531]: DEBUG oslo_concurrency.lockutils [None req-545f958f-5555-4fdc-8d01-3ef53b1af173 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Acquiring lock "f46d8a85-6ab7-436a-bc5e-40d0f80974e5" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.667436] env[63531]: DEBUG oslo_concurrency.lockutils [None req-545f958f-5555-4fdc-8d01-3ef53b1af173 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lock "f46d8a85-6ab7-436a-bc5e-40d0f80974e5" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.667675] env[63531]: INFO nova.compute.manager [None req-545f958f-5555-4fdc-8d01-3ef53b1af173 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Attaching volume c89ce39a-7d3d-468e-b6d6-7df150ed8817 to /dev/sdb [ 868.682120] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 868.682120] env[63531]: value = "task-1118014" [ 868.682120] env[63531]: _type = "Task" [ 868.682120] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.695076] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118014, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.718890] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4e8020e-0579-409f-b48b-8345ba80c23c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.730725] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64274a97-6f08-4787-9d1b-0b5c73903ab0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.751090] env[63531]: DEBUG nova.virt.block_device [None req-545f958f-5555-4fdc-8d01-3ef53b1af173 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Updating existing volume attachment record: 7b13f5e0-0e4b-4b4e-89a8-d2188c6be3bd {{(pid=63531) _volume_attach /opt/stack/nova/nova/virt/block_device.py:679}} [ 868.788594] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f2e31732-dbaf-4d48-baef-afe717b5167a tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.977991] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb65c966-577a-41e4-908e-eb5b0ae579ab {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.989994] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c90ee3af-2e5f-4675-b08f-2e4e5459fdd1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.051095] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35a3ef95-d09d-4f74-a86e-70dfed190d33 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.063367] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfb2eb7d-97be-4f81-aa6e-eca59257eb34 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.089450] env[63531]: DEBUG nova.compute.provider_tree [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 869.115903] env[63531]: INFO nova.compute.manager [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Took 39.98 seconds to build instance. [ 869.195416] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118014, 'name': CreateVM_Task} progress is 25%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.217779] env[63531]: DEBUG oslo_concurrency.lockutils [None req-36b75f90-354a-4910-9550-00922114a1d3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "2b0e4285-310a-439c-bc50-000a7c5ef7f9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.218169] env[63531]: DEBUG oslo_concurrency.lockutils [None req-36b75f90-354a-4910-9550-00922114a1d3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "2b0e4285-310a-439c-bc50-000a7c5ef7f9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.218401] env[63531]: DEBUG oslo_concurrency.lockutils [None req-36b75f90-354a-4910-9550-00922114a1d3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "2b0e4285-310a-439c-bc50-000a7c5ef7f9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.218587] env[63531]: DEBUG oslo_concurrency.lockutils [None req-36b75f90-354a-4910-9550-00922114a1d3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "2b0e4285-310a-439c-bc50-000a7c5ef7f9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.218775] env[63531]: DEBUG oslo_concurrency.lockutils [None req-36b75f90-354a-4910-9550-00922114a1d3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "2b0e4285-310a-439c-bc50-000a7c5ef7f9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.224928] env[63531]: INFO nova.compute.manager [None req-36b75f90-354a-4910-9550-00922114a1d3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Terminating instance [ 869.231173] env[63531]: DEBUG nova.compute.manager [None req-36b75f90-354a-4910-9550-00922114a1d3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 869.231173] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-36b75f90-354a-4910-9550-00922114a1d3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 869.232814] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7715b2d3-cf0e-4fff-ac93-0d5e4c417edb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.253761] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-36b75f90-354a-4910-9550-00922114a1d3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 869.256505] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9b408e7d-8792-409d-9d9e-bd0b5c1a8205 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.268120] env[63531]: DEBUG oslo_vmware.api [None req-36b75f90-354a-4910-9550-00922114a1d3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 869.268120] env[63531]: value = "task-1118018" [ 869.268120] env[63531]: _type = "Task" [ 869.268120] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.279961] env[63531]: DEBUG oslo_vmware.api [None req-36b75f90-354a-4910-9550-00922114a1d3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118018, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.289352] env[63531]: DEBUG nova.compute.manager [req-f0fb2524-2104-4f73-883b-515861d6b324 req-4526bc22-c6b6-4251-9644-ade426322f0d service nova] [instance: e95d3273-8216-47cc-95b6-99301366a827] Received event network-vif-plugged-a806c2a6-5a84-4114-aedb-8bb2da53a147 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 869.289589] env[63531]: DEBUG oslo_concurrency.lockutils [req-f0fb2524-2104-4f73-883b-515861d6b324 req-4526bc22-c6b6-4251-9644-ade426322f0d service nova] Acquiring lock "e95d3273-8216-47cc-95b6-99301366a827-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.290770] env[63531]: DEBUG oslo_concurrency.lockutils [req-f0fb2524-2104-4f73-883b-515861d6b324 req-4526bc22-c6b6-4251-9644-ade426322f0d service nova] Lock "e95d3273-8216-47cc-95b6-99301366a827-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.290770] env[63531]: DEBUG oslo_concurrency.lockutils [req-f0fb2524-2104-4f73-883b-515861d6b324 req-4526bc22-c6b6-4251-9644-ade426322f0d service nova] Lock "e95d3273-8216-47cc-95b6-99301366a827-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.290770] env[63531]: DEBUG nova.compute.manager [req-f0fb2524-2104-4f73-883b-515861d6b324 req-4526bc22-c6b6-4251-9644-ade426322f0d service nova] [instance: e95d3273-8216-47cc-95b6-99301366a827] No waiting events found dispatching network-vif-plugged-a806c2a6-5a84-4114-aedb-8bb2da53a147 {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 869.290770] env[63531]: WARNING nova.compute.manager [req-f0fb2524-2104-4f73-883b-515861d6b324 req-4526bc22-c6b6-4251-9644-ade426322f0d service nova] [instance: e95d3273-8216-47cc-95b6-99301366a827] Received unexpected event network-vif-plugged-a806c2a6-5a84-4114-aedb-8bb2da53a147 for instance with vm_state building and task_state spawning. [ 869.290770] env[63531]: DEBUG nova.compute.manager [req-f0fb2524-2104-4f73-883b-515861d6b324 req-4526bc22-c6b6-4251-9644-ade426322f0d service nova] [instance: e95d3273-8216-47cc-95b6-99301366a827] Received event network-changed-a806c2a6-5a84-4114-aedb-8bb2da53a147 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 869.290770] env[63531]: DEBUG nova.compute.manager [req-f0fb2524-2104-4f73-883b-515861d6b324 req-4526bc22-c6b6-4251-9644-ade426322f0d service nova] [instance: e95d3273-8216-47cc-95b6-99301366a827] Refreshing instance network info cache due to event network-changed-a806c2a6-5a84-4114-aedb-8bb2da53a147. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 869.291208] env[63531]: DEBUG oslo_concurrency.lockutils [req-f0fb2524-2104-4f73-883b-515861d6b324 req-4526bc22-c6b6-4251-9644-ade426322f0d service nova] Acquiring lock "refresh_cache-e95d3273-8216-47cc-95b6-99301366a827" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 869.291208] env[63531]: DEBUG oslo_concurrency.lockutils [req-f0fb2524-2104-4f73-883b-515861d6b324 req-4526bc22-c6b6-4251-9644-ade426322f0d service nova] Acquired lock "refresh_cache-e95d3273-8216-47cc-95b6-99301366a827" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.291385] env[63531]: DEBUG nova.network.neutron [req-f0fb2524-2104-4f73-883b-515861d6b324 req-4526bc22-c6b6-4251-9644-ade426322f0d service nova] [instance: e95d3273-8216-47cc-95b6-99301366a827] Refreshing network info cache for port a806c2a6-5a84-4114-aedb-8bb2da53a147 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 869.595913] env[63531]: DEBUG nova.scheduler.client.report [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 869.618208] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0855121-7260-4eb2-a315-ea560a91821f tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Lock "879924d4-b465-4102-a0e3-c7b2be7ef08b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 52.992s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.695962] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118014, 'name': CreateVM_Task} progress is 25%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.748423] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6a5546ec-a279-4010-b37e-f0e5bb9ca4fe tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Acquiring lock "db83adf4-6183-4330-b260-77d1f5daf899" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.749226] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6a5546ec-a279-4010-b37e-f0e5bb9ca4fe tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Lock "db83adf4-6183-4330-b260-77d1f5daf899" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.749226] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6a5546ec-a279-4010-b37e-f0e5bb9ca4fe tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Acquiring lock "db83adf4-6183-4330-b260-77d1f5daf899-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.749335] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6a5546ec-a279-4010-b37e-f0e5bb9ca4fe tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Lock "db83adf4-6183-4330-b260-77d1f5daf899-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.749431] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6a5546ec-a279-4010-b37e-f0e5bb9ca4fe tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Lock "db83adf4-6183-4330-b260-77d1f5daf899-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.752248] env[63531]: INFO nova.compute.manager [None req-6a5546ec-a279-4010-b37e-f0e5bb9ca4fe tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: db83adf4-6183-4330-b260-77d1f5daf899] Terminating instance [ 869.756692] env[63531]: DEBUG nova.compute.manager [None req-6a5546ec-a279-4010-b37e-f0e5bb9ca4fe tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: db83adf4-6183-4330-b260-77d1f5daf899] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 869.756935] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-6a5546ec-a279-4010-b37e-f0e5bb9ca4fe tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: db83adf4-6183-4330-b260-77d1f5daf899] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 869.759133] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1833824d-7a83-4ca7-85b4-96a303e20edc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.768122] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a5546ec-a279-4010-b37e-f0e5bb9ca4fe tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: db83adf4-6183-4330-b260-77d1f5daf899] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 869.768398] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-113ab186-f4ac-44a8-a012-a72631e0c95b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.782831] env[63531]: DEBUG oslo_vmware.api [None req-36b75f90-354a-4910-9550-00922114a1d3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118018, 'name': PowerOffVM_Task, 'duration_secs': 0.288635} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.784443] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-36b75f90-354a-4910-9550-00922114a1d3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 869.784629] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-36b75f90-354a-4910-9550-00922114a1d3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 869.788020] env[63531]: DEBUG oslo_vmware.api [None req-6a5546ec-a279-4010-b37e-f0e5bb9ca4fe tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Waiting for the task: (returnval){ [ 869.788020] env[63531]: value = "task-1118019" [ 869.788020] env[63531]: _type = "Task" [ 869.788020] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.788020] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4d550249-9af0-4277-907d-f7cd66a8f77d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.799640] env[63531]: DEBUG oslo_vmware.api [None req-6a5546ec-a279-4010-b37e-f0e5bb9ca4fe tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1118019, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.884786] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-36b75f90-354a-4910-9550-00922114a1d3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 869.885070] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-36b75f90-354a-4910-9550-00922114a1d3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Deleting contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 869.885326] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-36b75f90-354a-4910-9550-00922114a1d3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Deleting the datastore file [datastore2] 2b0e4285-310a-439c-bc50-000a7c5ef7f9 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 869.885662] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-638cbee5-7d3a-4701-b22e-2869ce35a4b1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.895655] env[63531]: DEBUG oslo_vmware.api [None req-36b75f90-354a-4910-9550-00922114a1d3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 869.895655] env[63531]: value = "task-1118021" [ 869.895655] env[63531]: _type = "Task" [ 869.895655] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.909361] env[63531]: DEBUG oslo_vmware.api [None req-36b75f90-354a-4910-9550-00922114a1d3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118021, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.103635] env[63531]: DEBUG oslo_vmware.rw_handles [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52a2690c-2c06-2f74-e0fe-59cf659a8252/disk-0.vmdk. {{(pid=63531) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 870.104497] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.578s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.104993] env[63531]: DEBUG nova.compute.manager [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 870.109653] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87ca56b1-7fe0-4221-9772-55ecbd0361cc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.112838] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a2338418-703e-4117-a6e8-2c9a448381e7 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.111s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.113121] env[63531]: DEBUG nova.objects.instance [None req-a2338418-703e-4117-a6e8-2c9a448381e7 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Lazy-loading 'resources' on Instance uuid 64fb5831-0789-482b-bb98-67b29868c4c9 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 870.121032] env[63531]: DEBUG oslo_vmware.rw_handles [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52a2690c-2c06-2f74-e0fe-59cf659a8252/disk-0.vmdk is in state: ready. {{(pid=63531) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 870.121132] env[63531]: ERROR oslo_vmware.rw_handles [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52a2690c-2c06-2f74-e0fe-59cf659a8252/disk-0.vmdk due to incomplete transfer. [ 870.121328] env[63531]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-f566a52a-0d3a-42d1-9f95-48b0bcd39f47 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.130592] env[63531]: DEBUG oslo_vmware.rw_handles [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52a2690c-2c06-2f74-e0fe-59cf659a8252/disk-0.vmdk. {{(pid=63531) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 870.130843] env[63531]: DEBUG nova.virt.vmwareapi.images [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Uploaded image d477826e-3e39-4ca7-bc24-998780166e1c to the Glance image server {{(pid=63531) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 870.132429] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Destroying the VM {{(pid=63531) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 870.133402] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-3f7ec374-fffa-422d-8977-106b56dd23c6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.145025] env[63531]: DEBUG oslo_vmware.api [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Waiting for the task: (returnval){ [ 870.145025] env[63531]: value = "task-1118022" [ 870.145025] env[63531]: _type = "Task" [ 870.145025] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.151816] env[63531]: DEBUG oslo_vmware.api [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': task-1118022, 'name': Destroy_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.181304] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8a9dc91a-31e6-4fb7-b47e-0b940db4e5dc tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Acquiring lock "e7fb00a3-0d2e-4f54-950d-337307112d7a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.181304] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8a9dc91a-31e6-4fb7-b47e-0b940db4e5dc tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Lock "e7fb00a3-0d2e-4f54-950d-337307112d7a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.181445] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8a9dc91a-31e6-4fb7-b47e-0b940db4e5dc tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Acquiring lock "e7fb00a3-0d2e-4f54-950d-337307112d7a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.181626] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8a9dc91a-31e6-4fb7-b47e-0b940db4e5dc tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Lock "e7fb00a3-0d2e-4f54-950d-337307112d7a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.181957] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8a9dc91a-31e6-4fb7-b47e-0b940db4e5dc tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Lock "e7fb00a3-0d2e-4f54-950d-337307112d7a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.184898] env[63531]: INFO nova.compute.manager [None req-8a9dc91a-31e6-4fb7-b47e-0b940db4e5dc tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Terminating instance [ 870.187048] env[63531]: DEBUG nova.compute.manager [None req-8a9dc91a-31e6-4fb7-b47e-0b940db4e5dc tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 870.187311] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-8a9dc91a-31e6-4fb7-b47e-0b940db4e5dc tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 870.191581] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c41fb97-404b-49a6-b18e-2d2888843bc9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.195220] env[63531]: DEBUG nova.network.neutron [req-f0fb2524-2104-4f73-883b-515861d6b324 req-4526bc22-c6b6-4251-9644-ade426322f0d service nova] [instance: e95d3273-8216-47cc-95b6-99301366a827] Updated VIF entry in instance network info cache for port a806c2a6-5a84-4114-aedb-8bb2da53a147. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 870.195697] env[63531]: DEBUG nova.network.neutron [req-f0fb2524-2104-4f73-883b-515861d6b324 req-4526bc22-c6b6-4251-9644-ade426322f0d service nova] [instance: e95d3273-8216-47cc-95b6-99301366a827] Updating instance_info_cache with network_info: [{"id": "a806c2a6-5a84-4114-aedb-8bb2da53a147", "address": "fa:16:3e:1e:bc:64", "network": {"id": "eac8786d-bc67-4916-8483-b53d9a25efb8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1604989690-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "abdae960d43b47bc95ab41f265d2c85b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db068f71-08cc-42d4-8ab6-17134c1585e5", "external-id": "nsx-vlan-transportzone-721", "segmentation_id": 721, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa806c2a6-5a", "ovs_interfaceid": "a806c2a6-5a84-4114-aedb-8bb2da53a147", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 870.208443] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a9dc91a-31e6-4fb7-b47e-0b940db4e5dc tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 870.211021] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118014, 'name': CreateVM_Task} progress is 25%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.211021] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c08fe19c-80ef-4fd4-8f6f-762a079b15d8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.217329] env[63531]: DEBUG oslo_vmware.api [None req-8a9dc91a-31e6-4fb7-b47e-0b940db4e5dc tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Waiting for the task: (returnval){ [ 870.217329] env[63531]: value = "task-1118023" [ 870.217329] env[63531]: _type = "Task" [ 870.217329] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.231019] env[63531]: DEBUG oslo_vmware.api [None req-8a9dc91a-31e6-4fb7-b47e-0b940db4e5dc tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1118023, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.237063] env[63531]: DEBUG oslo_concurrency.lockutils [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "a090135f-489d-457d-be36-ba6f61b71ab8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.237385] env[63531]: DEBUG oslo_concurrency.lockutils [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "a090135f-489d-457d-be36-ba6f61b71ab8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.299264] env[63531]: DEBUG oslo_vmware.api [None req-6a5546ec-a279-4010-b37e-f0e5bb9ca4fe tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1118019, 'name': PowerOffVM_Task, 'duration_secs': 0.259297} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.299614] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a5546ec-a279-4010-b37e-f0e5bb9ca4fe tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: db83adf4-6183-4330-b260-77d1f5daf899] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 870.299820] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-6a5546ec-a279-4010-b37e-f0e5bb9ca4fe tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: db83adf4-6183-4330-b260-77d1f5daf899] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 870.300123] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fe64c1f3-a723-4276-98cf-e3218cc48da0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.377604] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-6a5546ec-a279-4010-b37e-f0e5bb9ca4fe tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: db83adf4-6183-4330-b260-77d1f5daf899] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 870.378677] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-6a5546ec-a279-4010-b37e-f0e5bb9ca4fe tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: db83adf4-6183-4330-b260-77d1f5daf899] Deleting contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 870.378677] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a5546ec-a279-4010-b37e-f0e5bb9ca4fe tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Deleting the datastore file [datastore2] db83adf4-6183-4330-b260-77d1f5daf899 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 870.378808] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-912877b1-1c59-4d74-b8a9-14291b5215c6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.387038] env[63531]: DEBUG oslo_vmware.api [None req-6a5546ec-a279-4010-b37e-f0e5bb9ca4fe tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Waiting for the task: (returnval){ [ 870.387038] env[63531]: value = "task-1118025" [ 870.387038] env[63531]: _type = "Task" [ 870.387038] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.396578] env[63531]: DEBUG oslo_vmware.api [None req-6a5546ec-a279-4010-b37e-f0e5bb9ca4fe tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1118025, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.406535] env[63531]: DEBUG oslo_vmware.api [None req-36b75f90-354a-4910-9550-00922114a1d3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118021, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.228513} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.406713] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-36b75f90-354a-4910-9550-00922114a1d3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 870.407195] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-36b75f90-354a-4910-9550-00922114a1d3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Deleted contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 870.407285] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-36b75f90-354a-4910-9550-00922114a1d3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 870.407525] env[63531]: INFO nova.compute.manager [None req-36b75f90-354a-4910-9550-00922114a1d3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Took 1.18 seconds to destroy the instance on the hypervisor. [ 870.407848] env[63531]: DEBUG oslo.service.loopingcall [None req-36b75f90-354a-4910-9550-00922114a1d3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 870.408098] env[63531]: DEBUG nova.compute.manager [-] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 870.408312] env[63531]: DEBUG nova.network.neutron [-] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 870.615273] env[63531]: DEBUG nova.compute.utils [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 870.618443] env[63531]: DEBUG nova.compute.manager [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 870.618443] env[63531]: DEBUG nova.network.neutron [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 870.661682] env[63531]: DEBUG oslo_vmware.api [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': task-1118022, 'name': Destroy_Task, 'duration_secs': 0.394321} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.661965] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Destroyed the VM [ 870.662222] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Deleting Snapshot of the VM instance {{(pid=63531) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 870.662657] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-55cf0718-56ea-4787-94b3-26462b4067bc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.674959] env[63531]: DEBUG nova.policy [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1ea586d8e4f947d2902c130530d42466', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f9056486e3b640d0bcdb9369ed5d67ad', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 870.676996] env[63531]: DEBUG nova.compute.manager [None req-89c9bc13-a8c9-4b1a-9244-7179ff01bbf6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 870.679460] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92c42a5c-bef9-4a66-83f2-5294b7fc7caa {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.683440] env[63531]: DEBUG oslo_vmware.api [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Waiting for the task: (returnval){ [ 870.683440] env[63531]: value = "task-1118026" [ 870.683440] env[63531]: _type = "Task" [ 870.683440] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.699176] env[63531]: DEBUG oslo_concurrency.lockutils [req-f0fb2524-2104-4f73-883b-515861d6b324 req-4526bc22-c6b6-4251-9644-ade426322f0d service nova] Releasing lock "refresh_cache-e95d3273-8216-47cc-95b6-99301366a827" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.699874] env[63531]: DEBUG nova.compute.manager [req-f0fb2524-2104-4f73-883b-515861d6b324 req-4526bc22-c6b6-4251-9644-ade426322f0d service nova] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Received event network-vif-deleted-cbb5eb21-bc39-4608-820f-a77b3f5fdedf {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 870.708833] env[63531]: DEBUG oslo_vmware.api [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': task-1118026, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.708940] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118014, 'name': CreateVM_Task} progress is 25%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.728026] env[63531]: DEBUG oslo_vmware.api [None req-8a9dc91a-31e6-4fb7-b47e-0b940db4e5dc tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1118023, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.739821] env[63531]: DEBUG nova.compute.manager [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 870.899306] env[63531]: DEBUG oslo_vmware.api [None req-6a5546ec-a279-4010-b37e-f0e5bb9ca4fe tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1118025, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.176777} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.899580] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a5546ec-a279-4010-b37e-f0e5bb9ca4fe tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 870.899879] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-6a5546ec-a279-4010-b37e-f0e5bb9ca4fe tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: db83adf4-6183-4330-b260-77d1f5daf899] Deleted contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 870.900194] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-6a5546ec-a279-4010-b37e-f0e5bb9ca4fe tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: db83adf4-6183-4330-b260-77d1f5daf899] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 870.900749] env[63531]: INFO nova.compute.manager [None req-6a5546ec-a279-4010-b37e-f0e5bb9ca4fe tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: db83adf4-6183-4330-b260-77d1f5daf899] Took 1.14 seconds to destroy the instance on the hypervisor. [ 870.901016] env[63531]: DEBUG oslo.service.loopingcall [None req-6a5546ec-a279-4010-b37e-f0e5bb9ca4fe tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 870.901982] env[63531]: DEBUG nova.compute.manager [-] [instance: db83adf4-6183-4330-b260-77d1f5daf899] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 870.902089] env[63531]: DEBUG nova.network.neutron [-] [instance: db83adf4-6183-4330-b260-77d1f5daf899] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 871.068400] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5423ceb-d446-483b-8096-a745df8864cf {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.075768] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-733e7902-4adc-4fab-b03b-e00a75c2301f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.112767] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26b008d3-c09d-4f01-a128-958d272aac64 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.118510] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70a1dea7-645c-49d0-97cc-86f344a41768 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.122904] env[63531]: DEBUG nova.compute.manager [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 871.137240] env[63531]: DEBUG nova.compute.provider_tree [None req-a2338418-703e-4117-a6e8-2c9a448381e7 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 871.197558] env[63531]: DEBUG oslo_vmware.api [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': task-1118026, 'name': RemoveSnapshot_Task, 'duration_secs': 0.425749} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.198163] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Deleted Snapshot of the VM instance {{(pid=63531) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 871.198459] env[63531]: DEBUG nova.compute.manager [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 871.199728] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f97b19a-acc6-469c-a94a-9120b504f9a1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.205758] env[63531]: INFO nova.compute.manager [None req-89c9bc13-a8c9-4b1a-9244-7179ff01bbf6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] instance snapshotting [ 871.207100] env[63531]: DEBUG nova.objects.instance [None req-89c9bc13-a8c9-4b1a-9244-7179ff01bbf6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lazy-loading 'flavor' on Instance uuid 9a3fad46-a15a-451c-bdab-a3c8cc8add07 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 871.210107] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118014, 'name': CreateVM_Task} progress is 25%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.231247] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.231247] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.237410] env[63531]: DEBUG oslo_vmware.api [None req-8a9dc91a-31e6-4fb7-b47e-0b940db4e5dc tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1118023, 'name': PowerOffVM_Task, 'duration_secs': 0.799028} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.238072] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a9dc91a-31e6-4fb7-b47e-0b940db4e5dc tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 871.238322] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-8a9dc91a-31e6-4fb7-b47e-0b940db4e5dc tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 871.238635] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-16b38748-d51b-4953-9496-658710e204b0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.276974] env[63531]: DEBUG oslo_concurrency.lockutils [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.278483] env[63531]: DEBUG nova.network.neutron [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Successfully created port: 073adca7-1360-4703-9eb1-d86758bb8002 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 871.311452] env[63531]: DEBUG nova.network.neutron [-] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.382240] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-8a9dc91a-31e6-4fb7-b47e-0b940db4e5dc tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 871.382240] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-8a9dc91a-31e6-4fb7-b47e-0b940db4e5dc tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Deleting contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 871.382240] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-8a9dc91a-31e6-4fb7-b47e-0b940db4e5dc tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Deleting the datastore file [datastore2] e7fb00a3-0d2e-4f54-950d-337307112d7a {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 871.382240] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3354bf2c-1687-40ca-a3ac-38b94a7246a8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.392826] env[63531]: DEBUG oslo_vmware.api [None req-8a9dc91a-31e6-4fb7-b47e-0b940db4e5dc tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Waiting for the task: (returnval){ [ 871.392826] env[63531]: value = "task-1118029" [ 871.392826] env[63531]: _type = "Task" [ 871.392826] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.406696] env[63531]: DEBUG oslo_vmware.api [None req-8a9dc91a-31e6-4fb7-b47e-0b940db4e5dc tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1118029, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.477165] env[63531]: DEBUG nova.compute.manager [req-cd344d86-a3a9-4b31-9d4a-e062ea855e96 req-6b3e52a9-0a8e-4c3a-af98-0509a6f53631 service nova] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Received event network-vif-deleted-13365dfb-f81c-4e56-bf30-0a09dec4f561 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 871.639219] env[63531]: DEBUG nova.scheduler.client.report [None req-a2338418-703e-4117-a6e8-2c9a448381e7 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 871.701579] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118014, 'name': CreateVM_Task} progress is 25%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.713379] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc9d584f-2668-4ef6-b336-c77f7a1c4717 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.733608] env[63531]: INFO nova.compute.manager [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Shelve offloading [ 871.737239] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cb2deb7-b551-498c-8a46-153fae80cf57 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.739981] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 871.741341] env[63531]: DEBUG nova.compute.manager [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 871.747156] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-470d9595-9343-442f-ab6b-04797117e693 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.755761] env[63531]: DEBUG oslo_vmware.api [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Waiting for the task: (returnval){ [ 871.755761] env[63531]: value = "task-1118030" [ 871.755761] env[63531]: _type = "Task" [ 871.755761] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.765977] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] VM already powered off {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 871.766311] env[63531]: DEBUG nova.compute.manager [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 871.767326] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1175953b-86a7-4162-b6d0-fde0d6d33f23 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.778796] env[63531]: DEBUG oslo_concurrency.lockutils [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Acquiring lock "refresh_cache-2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.779083] env[63531]: DEBUG oslo_concurrency.lockutils [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Acquired lock "refresh_cache-2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.779185] env[63531]: DEBUG nova.network.neutron [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 871.816350] env[63531]: INFO nova.compute.manager [-] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Took 1.41 seconds to deallocate network for instance. [ 871.900224] env[63531]: DEBUG nova.network.neutron [-] [instance: db83adf4-6183-4330-b260-77d1f5daf899] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.913022] env[63531]: DEBUG oslo_vmware.api [None req-8a9dc91a-31e6-4fb7-b47e-0b940db4e5dc tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Task: {'id': task-1118029, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.189046} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.913022] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-8a9dc91a-31e6-4fb7-b47e-0b940db4e5dc tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 871.913022] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-8a9dc91a-31e6-4fb7-b47e-0b940db4e5dc tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Deleted contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 871.913022] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-8a9dc91a-31e6-4fb7-b47e-0b940db4e5dc tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 871.913022] env[63531]: INFO nova.compute.manager [None req-8a9dc91a-31e6-4fb7-b47e-0b940db4e5dc tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Took 1.72 seconds to destroy the instance on the hypervisor. [ 871.913022] env[63531]: DEBUG oslo.service.loopingcall [None req-8a9dc91a-31e6-4fb7-b47e-0b940db4e5dc tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 871.913022] env[63531]: DEBUG nova.compute.manager [-] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 871.913022] env[63531]: DEBUG nova.network.neutron [-] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 871.964019] env[63531]: INFO nova.compute.manager [None req-5ca07bd6-10c2-4158-87cf-a11b8952cc25 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Rescuing [ 871.964019] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5ca07bd6-10c2-4158-87cf-a11b8952cc25 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquiring lock "refresh_cache-879924d4-b465-4102-a0e3-c7b2be7ef08b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.964019] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5ca07bd6-10c2-4158-87cf-a11b8952cc25 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquired lock "refresh_cache-879924d4-b465-4102-a0e3-c7b2be7ef08b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.964019] env[63531]: DEBUG nova.network.neutron [None req-5ca07bd6-10c2-4158-87cf-a11b8952cc25 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 872.133941] env[63531]: DEBUG nova.compute.manager [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 872.148208] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a2338418-703e-4117-a6e8-2c9a448381e7 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.034s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.149425] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1654cc47-4170-4ce7-ab39-6549bb1aa779 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.501s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.149757] env[63531]: DEBUG nova.objects.instance [None req-1654cc47-4170-4ce7-ab39-6549bb1aa779 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Lazy-loading 'resources' on Instance uuid 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 872.174391] env[63531]: DEBUG nova.virt.hardware [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 872.174765] env[63531]: DEBUG nova.virt.hardware [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 872.174932] env[63531]: DEBUG nova.virt.hardware [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 872.175107] env[63531]: DEBUG nova.virt.hardware [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 872.175302] env[63531]: DEBUG nova.virt.hardware [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 872.175469] env[63531]: DEBUG nova.virt.hardware [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 872.175693] env[63531]: DEBUG nova.virt.hardware [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 872.175859] env[63531]: DEBUG nova.virt.hardware [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 872.176042] env[63531]: DEBUG nova.virt.hardware [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 872.176217] env[63531]: DEBUG nova.virt.hardware [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 872.176397] env[63531]: DEBUG nova.virt.hardware [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 872.177740] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6184c03-85be-4bdb-bf5e-dd19d657299f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.182448] env[63531]: INFO nova.scheduler.client.report [None req-a2338418-703e-4117-a6e8-2c9a448381e7 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Deleted allocations for instance 64fb5831-0789-482b-bb98-67b29868c4c9 [ 872.194692] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b232d566-b3cd-499b-b3ec-d9902919ee0e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.217984] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118014, 'name': CreateVM_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.257346] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-89c9bc13-a8c9-4b1a-9244-7179ff01bbf6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Creating Snapshot of the VM instance {{(pid=63531) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 872.257666] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-3038e656-737b-4d43-a16a-4c41262fcf6d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.261703] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.266547] env[63531]: DEBUG oslo_vmware.api [None req-89c9bc13-a8c9-4b1a-9244-7179ff01bbf6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 872.266547] env[63531]: value = "task-1118031" [ 872.266547] env[63531]: _type = "Task" [ 872.266547] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.275228] env[63531]: DEBUG oslo_vmware.api [None req-89c9bc13-a8c9-4b1a-9244-7179ff01bbf6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118031, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.322010] env[63531]: DEBUG oslo_concurrency.lockutils [None req-36b75f90-354a-4910-9550-00922114a1d3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.404142] env[63531]: INFO nova.compute.manager [-] [instance: db83adf4-6183-4330-b260-77d1f5daf899] Took 1.50 seconds to deallocate network for instance. [ 872.529980] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "5ad9fc53-104b-4a90-801a-bd1e8f12c1d7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.529980] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "5ad9fc53-104b-4a90-801a-bd1e8f12c1d7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.573115] env[63531]: DEBUG nova.network.neutron [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Updating instance_info_cache with network_info: [{"id": "2d27a8a5-c288-44fe-ac09-0d3b488386d6", "address": "fa:16:3e:91:1c:62", "network": {"id": "5b8b7134-f123-40b8-9134-5cd072e8ad01", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1232565840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.165", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef3b56ce3f3a4155bd09642f5c48ea6f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2d27a8a5-c2", "ovs_interfaceid": "2d27a8a5-c288-44fe-ac09-0d3b488386d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.698039] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a2338418-703e-4117-a6e8-2c9a448381e7 tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Lock "64fb5831-0789-482b-bb98-67b29868c4c9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.971s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.700326] env[63531]: DEBUG nova.network.neutron [-] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.714038] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118014, 'name': CreateVM_Task, 'duration_secs': 3.537124} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.714038] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e95d3273-8216-47cc-95b6-99301366a827] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 872.714038] env[63531]: DEBUG oslo_concurrency.lockutils [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 872.714358] env[63531]: DEBUG oslo_concurrency.lockutils [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.714649] env[63531]: DEBUG oslo_concurrency.lockutils [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 872.718884] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4e9826ee-86db-41f6-aa75-5bcfa2a3c2d4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.729753] env[63531]: DEBUG oslo_vmware.api [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 872.729753] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]528cd501-332d-e6ad-3e3e-f57507c85ad5" [ 872.729753] env[63531]: _type = "Task" [ 872.729753] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.739977] env[63531]: DEBUG oslo_vmware.api [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]528cd501-332d-e6ad-3e3e-f57507c85ad5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.780202] env[63531]: DEBUG oslo_vmware.api [None req-89c9bc13-a8c9-4b1a-9244-7179ff01bbf6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118031, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.907837] env[63531]: DEBUG nova.network.neutron [None req-5ca07bd6-10c2-4158-87cf-a11b8952cc25 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Updating instance_info_cache with network_info: [{"id": "d5d578b0-1f4a-4a46-8fac-1e625d48c2bc", "address": "fa:16:3e:ca:1a:aa", "network": {"id": "913045d9-e90b-4366-987c-7706b73a6147", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1696510389-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f71c7db0a3614995becc25d9a8dce787", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c297fe21-cd0b-4226-813b-a65d2358d034", "external-id": "nsx-vlan-transportzone-98", "segmentation_id": 98, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5d578b0-1f", "ovs_interfaceid": "d5d578b0-1f4a-4a46-8fac-1e625d48c2bc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.912896] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6a5546ec-a279-4010-b37e-f0e5bb9ca4fe tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.952176] env[63531]: DEBUG nova.network.neutron [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Successfully updated port: 073adca7-1360-4703-9eb1-d86758bb8002 {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 873.033750] env[63531]: DEBUG nova.compute.manager [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 873.047083] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53a4385c-aa91-4e00-9223-152dc22c8b7c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.055348] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfe30bd8-bd58-4c8c-9446-85b0f09a136f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.085667] env[63531]: DEBUG oslo_concurrency.lockutils [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Releasing lock "refresh_cache-2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.089208] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faa1396a-a447-4cb3-9901-5cadbbd6bed0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.102796] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d85dd8fa-fe7f-4fab-bd1d-59a12f94aec8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.119012] env[63531]: DEBUG nova.compute.provider_tree [None req-1654cc47-4170-4ce7-ab39-6549bb1aa779 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 873.206665] env[63531]: INFO nova.compute.manager [-] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Took 1.29 seconds to deallocate network for instance. [ 873.243452] env[63531]: DEBUG oslo_vmware.api [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]528cd501-332d-e6ad-3e3e-f57507c85ad5, 'name': SearchDatastore_Task, 'duration_secs': 0.010966} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.245941] env[63531]: DEBUG oslo_concurrency.lockutils [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.245941] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 873.245941] env[63531]: DEBUG oslo_concurrency.lockutils [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.245941] env[63531]: DEBUG oslo_concurrency.lockutils [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.245941] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 873.245941] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-574fabd3-45d9-4422-a0b0-84cee6c6b134 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.254869] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 873.258262] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 873.258262] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-449aa191-5436-49fb-a259-b31612aa4021 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.268253] env[63531]: DEBUG oslo_vmware.api [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 873.268253] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5204a2b4-6303-76b4-b9ce-bfd6cb41c788" [ 873.268253] env[63531]: _type = "Task" [ 873.268253] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.279693] env[63531]: DEBUG oslo_vmware.api [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5204a2b4-6303-76b4-b9ce-bfd6cb41c788, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.283451] env[63531]: DEBUG oslo_vmware.api [None req-89c9bc13-a8c9-4b1a-9244-7179ff01bbf6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118031, 'name': CreateSnapshot_Task, 'duration_secs': 0.697398} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.283511] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-89c9bc13-a8c9-4b1a-9244-7179ff01bbf6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Created Snapshot of the VM instance {{(pid=63531) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 873.284344] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-618d3d48-4f23-4242-835e-bae9ec793c40 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.411171] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5ca07bd6-10c2-4158-87cf-a11b8952cc25 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Releasing lock "refresh_cache-879924d4-b465-4102-a0e3-c7b2be7ef08b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.454772] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquiring lock "refresh_cache-37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.454943] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquired lock "refresh_cache-37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.455362] env[63531]: DEBUG nova.network.neutron [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 873.506594] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 873.507725] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd353dbc-4094-4d9a-8809-ef6c7d864dac {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.516640] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 873.516915] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c65a5652-5ed7-4712-bb5b-6e3322476867 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.552838] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.602071] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 873.602071] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 873.602071] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Deleting the datastore file [datastore1] 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 873.602071] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9b48e7ff-5e9d-453f-b0d5-9efa1c95b810 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.608056] env[63531]: DEBUG oslo_vmware.api [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Waiting for the task: (returnval){ [ 873.608056] env[63531]: value = "task-1118033" [ 873.608056] env[63531]: _type = "Task" [ 873.608056] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.622451] env[63531]: DEBUG nova.scheduler.client.report [None req-1654cc47-4170-4ce7-ab39-6549bb1aa779 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 873.626388] env[63531]: DEBUG oslo_vmware.api [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': task-1118033, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.713965] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8a9dc91a-31e6-4fb7-b47e-0b940db4e5dc tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.717237] env[63531]: DEBUG nova.compute.manager [req-a7898dcb-d96d-4ec9-8d99-f5570e24c394 req-72aa5472-4e74-43ed-a6b5-dc8a9621095c service nova] [instance: db83adf4-6183-4330-b260-77d1f5daf899] Received event network-vif-deleted-895f47c1-21ad-45cc-b1d4-e7a8fa7398cb {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 873.717511] env[63531]: DEBUG nova.compute.manager [req-a7898dcb-d96d-4ec9-8d99-f5570e24c394 req-72aa5472-4e74-43ed-a6b5-dc8a9621095c service nova] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Received event network-vif-deleted-6102475b-8e43-4b77-85c3-5b574383c06d {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 873.717746] env[63531]: DEBUG nova.compute.manager [req-a7898dcb-d96d-4ec9-8d99-f5570e24c394 req-72aa5472-4e74-43ed-a6b5-dc8a9621095c service nova] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Received event network-vif-plugged-073adca7-1360-4703-9eb1-d86758bb8002 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 873.717969] env[63531]: DEBUG oslo_concurrency.lockutils [req-a7898dcb-d96d-4ec9-8d99-f5570e24c394 req-72aa5472-4e74-43ed-a6b5-dc8a9621095c service nova] Acquiring lock "37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.718239] env[63531]: DEBUG oslo_concurrency.lockutils [req-a7898dcb-d96d-4ec9-8d99-f5570e24c394 req-72aa5472-4e74-43ed-a6b5-dc8a9621095c service nova] Lock "37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.718477] env[63531]: DEBUG oslo_concurrency.lockutils [req-a7898dcb-d96d-4ec9-8d99-f5570e24c394 req-72aa5472-4e74-43ed-a6b5-dc8a9621095c service nova] Lock "37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.718768] env[63531]: DEBUG nova.compute.manager [req-a7898dcb-d96d-4ec9-8d99-f5570e24c394 req-72aa5472-4e74-43ed-a6b5-dc8a9621095c service nova] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] No waiting events found dispatching network-vif-plugged-073adca7-1360-4703-9eb1-d86758bb8002 {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 873.719015] env[63531]: WARNING nova.compute.manager [req-a7898dcb-d96d-4ec9-8d99-f5570e24c394 req-72aa5472-4e74-43ed-a6b5-dc8a9621095c service nova] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Received unexpected event network-vif-plugged-073adca7-1360-4703-9eb1-d86758bb8002 for instance with vm_state building and task_state spawning. [ 873.719307] env[63531]: DEBUG nova.compute.manager [req-a7898dcb-d96d-4ec9-8d99-f5570e24c394 req-72aa5472-4e74-43ed-a6b5-dc8a9621095c service nova] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Received event network-changed-073adca7-1360-4703-9eb1-d86758bb8002 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 873.719572] env[63531]: DEBUG nova.compute.manager [req-a7898dcb-d96d-4ec9-8d99-f5570e24c394 req-72aa5472-4e74-43ed-a6b5-dc8a9621095c service nova] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Refreshing instance network info cache due to event network-changed-073adca7-1360-4703-9eb1-d86758bb8002. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 873.719838] env[63531]: DEBUG oslo_concurrency.lockutils [req-a7898dcb-d96d-4ec9-8d99-f5570e24c394 req-72aa5472-4e74-43ed-a6b5-dc8a9621095c service nova] Acquiring lock "refresh_cache-37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.779438] env[63531]: DEBUG oslo_vmware.api [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5204a2b4-6303-76b4-b9ce-bfd6cb41c788, 'name': SearchDatastore_Task, 'duration_secs': 0.016253} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.780252] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-22c22a30-e54b-4315-9a13-736d3b64fe45 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.786344] env[63531]: DEBUG oslo_vmware.api [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 873.786344] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]528e1b14-6e0d-2c55-7334-bc862ee1a320" [ 873.786344] env[63531]: _type = "Task" [ 873.786344] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.794561] env[63531]: DEBUG oslo_concurrency.lockutils [None req-29b04bcd-cd4b-4a53-9b9f-4a106df45c8e tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Acquiring lock "2998fb3d-6677-4fb3-952a-c74d8c6b7b9f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.794805] env[63531]: DEBUG oslo_concurrency.lockutils [None req-29b04bcd-cd4b-4a53-9b9f-4a106df45c8e tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Lock "2998fb3d-6677-4fb3-952a-c74d8c6b7b9f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.795013] env[63531]: DEBUG oslo_concurrency.lockutils [None req-29b04bcd-cd4b-4a53-9b9f-4a106df45c8e tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Acquiring lock "2998fb3d-6677-4fb3-952a-c74d8c6b7b9f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.795207] env[63531]: DEBUG oslo_concurrency.lockutils [None req-29b04bcd-cd4b-4a53-9b9f-4a106df45c8e tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Lock "2998fb3d-6677-4fb3-952a-c74d8c6b7b9f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.795378] env[63531]: DEBUG oslo_concurrency.lockutils [None req-29b04bcd-cd4b-4a53-9b9f-4a106df45c8e tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Lock "2998fb3d-6677-4fb3-952a-c74d8c6b7b9f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.796973] env[63531]: DEBUG oslo_vmware.api [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]528e1b14-6e0d-2c55-7334-bc862ee1a320, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.797459] env[63531]: INFO nova.compute.manager [None req-29b04bcd-cd4b-4a53-9b9f-4a106df45c8e tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Terminating instance [ 873.799332] env[63531]: DEBUG nova.compute.manager [None req-29b04bcd-cd4b-4a53-9b9f-4a106df45c8e tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 873.800027] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-29b04bcd-cd4b-4a53-9b9f-4a106df45c8e tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 873.800410] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94512604-9e77-41b7-bd6e-293dd269dd94 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.809037] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-89c9bc13-a8c9-4b1a-9244-7179ff01bbf6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Creating linked-clone VM from snapshot {{(pid=63531) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 873.809287] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-974e6d78-780d-43f9-82f1-1bc96e950c32 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.813428] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-545f958f-5555-4fdc-8d01-3ef53b1af173 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Volume attach. Driver type: vmdk {{(pid=63531) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 873.813639] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-545f958f-5555-4fdc-8d01-3ef53b1af173 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244725', 'volume_id': 'c89ce39a-7d3d-468e-b6d6-7df150ed8817', 'name': 'volume-c89ce39a-7d3d-468e-b6d6-7df150ed8817', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f46d8a85-6ab7-436a-bc5e-40d0f80974e5', 'attached_at': '', 'detached_at': '', 'volume_id': 'c89ce39a-7d3d-468e-b6d6-7df150ed8817', 'serial': 'c89ce39a-7d3d-468e-b6d6-7df150ed8817'} {{(pid=63531) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 873.814369] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33e372c0-c478-448e-b3e6-a31331357606 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.818856] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-29b04bcd-cd4b-4a53-9b9f-4a106df45c8e tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 873.819426] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b1403c79-60fa-4f96-a8cc-6610b81f462b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.841809] env[63531]: DEBUG oslo_vmware.api [None req-89c9bc13-a8c9-4b1a-9244-7179ff01bbf6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 873.841809] env[63531]: value = "task-1118034" [ 873.841809] env[63531]: _type = "Task" [ 873.841809] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.847024] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-301ea774-224a-4173-8cba-4a1b03b6418c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.852381] env[63531]: DEBUG oslo_vmware.api [None req-29b04bcd-cd4b-4a53-9b9f-4a106df45c8e tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Waiting for the task: (returnval){ [ 873.852381] env[63531]: value = "task-1118035" [ 873.852381] env[63531]: _type = "Task" [ 873.852381] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.881975] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-545f958f-5555-4fdc-8d01-3ef53b1af173 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Reconfiguring VM instance instance-0000003c to attach disk [datastore1] volume-c89ce39a-7d3d-468e-b6d6-7df150ed8817/volume-c89ce39a-7d3d-468e-b6d6-7df150ed8817.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 873.882649] env[63531]: DEBUG oslo_vmware.api [None req-89c9bc13-a8c9-4b1a-9244-7179ff01bbf6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118034, 'name': CloneVM_Task} progress is 11%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.883521] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e3a56e56-717d-4293-a53e-5ad03b118928 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.902019] env[63531]: DEBUG oslo_vmware.api [None req-29b04bcd-cd4b-4a53-9b9f-4a106df45c8e tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1118035, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.909349] env[63531]: DEBUG oslo_vmware.api [None req-545f958f-5555-4fdc-8d01-3ef53b1af173 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Waiting for the task: (returnval){ [ 873.909349] env[63531]: value = "task-1118036" [ 873.909349] env[63531]: _type = "Task" [ 873.909349] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.937230] env[63531]: DEBUG oslo_vmware.api [None req-545f958f-5555-4fdc-8d01-3ef53b1af173 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118036, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.947980] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ca07bd6-10c2-4158-87cf-a11b8952cc25 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 873.948577] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9d6bbd28-088f-4cb9-abe3-a863cc72ab70 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.956861] env[63531]: DEBUG oslo_vmware.api [None req-5ca07bd6-10c2-4158-87cf-a11b8952cc25 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 873.956861] env[63531]: value = "task-1118037" [ 873.956861] env[63531]: _type = "Task" [ 873.956861] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.969192] env[63531]: DEBUG oslo_vmware.api [None req-5ca07bd6-10c2-4158-87cf-a11b8952cc25 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118037, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.997718] env[63531]: DEBUG nova.network.neutron [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 874.119786] env[63531]: DEBUG oslo_vmware.api [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': task-1118033, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.147857} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.120412] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 874.120412] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 874.120557] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 874.130547] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1654cc47-4170-4ce7-ab39-6549bb1aa779 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.981s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.132545] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 14.697s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.132696] env[63531]: DEBUG nova.objects.instance [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63531) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 874.143301] env[63531]: INFO nova.scheduler.client.report [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Deleted allocations for instance 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f [ 874.162021] env[63531]: INFO nova.scheduler.client.report [None req-1654cc47-4170-4ce7-ab39-6549bb1aa779 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Deleted allocations for instance 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d [ 874.175137] env[63531]: DEBUG nova.network.neutron [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Updating instance_info_cache with network_info: [{"id": "073adca7-1360-4703-9eb1-d86758bb8002", "address": "fa:16:3e:9b:a3:5c", "network": {"id": "95ed2c7d-3d1a-4736-8e90-104002f5c40b", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-73549344-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9056486e3b640d0bcdb9369ed5d67ad", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04e15990-16e1-4cb2-b0f0-06c362e68c5e", "external-id": "nsx-vlan-transportzone-555", "segmentation_id": 555, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap073adca7-13", "ovs_interfaceid": "073adca7-1360-4703-9eb1-d86758bb8002", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.299721] env[63531]: DEBUG oslo_vmware.api [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]528e1b14-6e0d-2c55-7334-bc862ee1a320, 'name': SearchDatastore_Task, 'duration_secs': 0.010191} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.299721] env[63531]: DEBUG oslo_concurrency.lockutils [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.299721] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] e95d3273-8216-47cc-95b6-99301366a827/e95d3273-8216-47cc-95b6-99301366a827.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 874.300204] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-14e32939-5ee9-44fe-b962-dd7ee7c2e179 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.309512] env[63531]: DEBUG oslo_vmware.api [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 874.309512] env[63531]: value = "task-1118038" [ 874.309512] env[63531]: _type = "Task" [ 874.309512] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.318670] env[63531]: DEBUG oslo_vmware.api [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118038, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.359407] env[63531]: DEBUG oslo_vmware.api [None req-89c9bc13-a8c9-4b1a-9244-7179ff01bbf6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118034, 'name': CloneVM_Task} progress is 94%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.369016] env[63531]: DEBUG oslo_vmware.api [None req-29b04bcd-cd4b-4a53-9b9f-4a106df45c8e tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1118035, 'name': PowerOffVM_Task, 'duration_secs': 0.430432} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.369344] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-29b04bcd-cd4b-4a53-9b9f-4a106df45c8e tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 874.369705] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-29b04bcd-cd4b-4a53-9b9f-4a106df45c8e tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 874.369937] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-89f8e222-1fad-4597-8ca6-844e04bd1529 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.423244] env[63531]: DEBUG oslo_vmware.api [None req-545f958f-5555-4fdc-8d01-3ef53b1af173 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118036, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.468235] env[63531]: DEBUG oslo_vmware.api [None req-5ca07bd6-10c2-4158-87cf-a11b8952cc25 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118037, 'name': PowerOffVM_Task, 'duration_secs': 0.23817} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.470795] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ca07bd6-10c2-4158-87cf-a11b8952cc25 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 874.470795] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-29b04bcd-cd4b-4a53-9b9f-4a106df45c8e tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 874.470795] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-29b04bcd-cd4b-4a53-9b9f-4a106df45c8e tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 874.470795] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-29b04bcd-cd4b-4a53-9b9f-4a106df45c8e tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Deleting the datastore file [datastore1] 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 874.471243] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84349d28-50c1-4e02-9e31-fa3f17d3034d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.474022] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3a3cbe32-e75b-4221-8819-9b8ba9590c3c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.495987] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f92d6d2c-7931-4593-9675-5b0ec5fa35f6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.499282] env[63531]: DEBUG oslo_vmware.api [None req-29b04bcd-cd4b-4a53-9b9f-4a106df45c8e tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Waiting for the task: (returnval){ [ 874.499282] env[63531]: value = "task-1118040" [ 874.499282] env[63531]: _type = "Task" [ 874.499282] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.513230] env[63531]: DEBUG oslo_vmware.api [None req-29b04bcd-cd4b-4a53-9b9f-4a106df45c8e tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1118040, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.535069] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ca07bd6-10c2-4158-87cf-a11b8952cc25 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 874.535430] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-58a6a0c5-036e-4489-b689-bcacf654566e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.544530] env[63531]: DEBUG oslo_vmware.api [None req-5ca07bd6-10c2-4158-87cf-a11b8952cc25 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 874.544530] env[63531]: value = "task-1118041" [ 874.544530] env[63531]: _type = "Task" [ 874.544530] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.554240] env[63531]: DEBUG oslo_vmware.api [None req-5ca07bd6-10c2-4158-87cf-a11b8952cc25 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118041, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.649336] env[63531]: DEBUG oslo_concurrency.lockutils [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.678228] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1654cc47-4170-4ce7-ab39-6549bb1aa779 tempest-ServerMetadataNegativeTestJSON-806453667 tempest-ServerMetadataNegativeTestJSON-806453667-project-member] Lock "56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.301s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.679608] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Releasing lock "refresh_cache-37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.679958] env[63531]: DEBUG nova.compute.manager [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Instance network_info: |[{"id": "073adca7-1360-4703-9eb1-d86758bb8002", "address": "fa:16:3e:9b:a3:5c", "network": {"id": "95ed2c7d-3d1a-4736-8e90-104002f5c40b", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-73549344-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9056486e3b640d0bcdb9369ed5d67ad", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04e15990-16e1-4cb2-b0f0-06c362e68c5e", "external-id": "nsx-vlan-transportzone-555", "segmentation_id": 555, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap073adca7-13", "ovs_interfaceid": "073adca7-1360-4703-9eb1-d86758bb8002", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 874.680519] env[63531]: DEBUG oslo_concurrency.lockutils [req-a7898dcb-d96d-4ec9-8d99-f5570e24c394 req-72aa5472-4e74-43ed-a6b5-dc8a9621095c service nova] Acquired lock "refresh_cache-37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.680741] env[63531]: DEBUG nova.network.neutron [req-a7898dcb-d96d-4ec9-8d99-f5570e24c394 req-72aa5472-4e74-43ed-a6b5-dc8a9621095c service nova] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Refreshing network info cache for port 073adca7-1360-4703-9eb1-d86758bb8002 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 874.682227] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9b:a3:5c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '04e15990-16e1-4cb2-b0f0-06c362e68c5e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '073adca7-1360-4703-9eb1-d86758bb8002', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 874.691784] env[63531]: DEBUG oslo.service.loopingcall [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 874.693050] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 874.693466] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ae96944d-63ed-4660-b6e6-4f15e913b666 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.720597] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 874.720597] env[63531]: value = "task-1118042" [ 874.720597] env[63531]: _type = "Task" [ 874.720597] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.731768] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118042, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.821326] env[63531]: DEBUG oslo_vmware.api [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118038, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.50725} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.822022] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] e95d3273-8216-47cc-95b6-99301366a827/e95d3273-8216-47cc-95b6-99301366a827.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 874.822022] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 874.822254] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9b25170e-b320-444e-b20b-22978e59136a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.829424] env[63531]: DEBUG oslo_vmware.api [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 874.829424] env[63531]: value = "task-1118043" [ 874.829424] env[63531]: _type = "Task" [ 874.829424] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.838819] env[63531]: DEBUG oslo_vmware.api [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118043, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.855335] env[63531]: DEBUG oslo_vmware.api [None req-89c9bc13-a8c9-4b1a-9244-7179ff01bbf6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118034, 'name': CloneVM_Task} progress is 95%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.922196] env[63531]: DEBUG oslo_vmware.api [None req-545f958f-5555-4fdc-8d01-3ef53b1af173 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118036, 'name': ReconfigVM_Task, 'duration_secs': 0.540925} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.923807] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-545f958f-5555-4fdc-8d01-3ef53b1af173 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Reconfigured VM instance instance-0000003c to attach disk [datastore1] volume-c89ce39a-7d3d-468e-b6d6-7df150ed8817/volume-c89ce39a-7d3d-468e-b6d6-7df150ed8817.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 874.928505] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-143fa7ce-738a-4635-beaa-ee263e52e49b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.946231] env[63531]: DEBUG oslo_vmware.api [None req-545f958f-5555-4fdc-8d01-3ef53b1af173 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Waiting for the task: (returnval){ [ 874.946231] env[63531]: value = "task-1118044" [ 874.946231] env[63531]: _type = "Task" [ 874.946231] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.956128] env[63531]: DEBUG oslo_vmware.api [None req-545f958f-5555-4fdc-8d01-3ef53b1af173 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118044, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.009026] env[63531]: DEBUG oslo_vmware.api [None req-29b04bcd-cd4b-4a53-9b9f-4a106df45c8e tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Task: {'id': task-1118040, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.444927} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.009433] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-29b04bcd-cd4b-4a53-9b9f-4a106df45c8e tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 875.009770] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-29b04bcd-cd4b-4a53-9b9f-4a106df45c8e tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 875.010092] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-29b04bcd-cd4b-4a53-9b9f-4a106df45c8e tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 875.010374] env[63531]: INFO nova.compute.manager [None req-29b04bcd-cd4b-4a53-9b9f-4a106df45c8e tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Took 1.21 seconds to destroy the instance on the hypervisor. [ 875.011595] env[63531]: DEBUG oslo.service.loopingcall [None req-29b04bcd-cd4b-4a53-9b9f-4a106df45c8e tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 875.011595] env[63531]: DEBUG nova.compute.manager [-] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 875.011595] env[63531]: DEBUG nova.network.neutron [-] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 875.058525] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ca07bd6-10c2-4158-87cf-a11b8952cc25 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] VM already powered off {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 875.058776] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5ca07bd6-10c2-4158-87cf-a11b8952cc25 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 875.059041] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5ca07bd6-10c2-4158-87cf-a11b8952cc25 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.059204] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5ca07bd6-10c2-4158-87cf-a11b8952cc25 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.059464] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ca07bd6-10c2-4158-87cf-a11b8952cc25 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 875.059774] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fa733814-67bb-4e80-a684-e1b168fb643a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.069038] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ca07bd6-10c2-4158-87cf-a11b8952cc25 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 875.069245] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5ca07bd6-10c2-4158-87cf-a11b8952cc25 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 875.069959] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8475da55-9d78-483a-8707-800bb7c60ab3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.075923] env[63531]: DEBUG oslo_vmware.api [None req-5ca07bd6-10c2-4158-87cf-a11b8952cc25 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 875.075923] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]529d5c00-87c7-cc73-7f7d-b569cae2e282" [ 875.075923] env[63531]: _type = "Task" [ 875.075923] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.085907] env[63531]: DEBUG oslo_vmware.api [None req-5ca07bd6-10c2-4158-87cf-a11b8952cc25 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]529d5c00-87c7-cc73-7f7d-b569cae2e282, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.146896] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b47f1569-09bd-4b68-b790-003bb6e4132c tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.014s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.148133] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1b340bf7-08de-4d21-a861-c6e587976b10 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.636s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.148420] env[63531]: DEBUG nova.objects.instance [None req-1b340bf7-08de-4d21-a861-c6e587976b10 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Lazy-loading 'resources' on Instance uuid 09feb12a-da9f-4bb7-959d-84c5c7c58e99 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 875.234760] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118042, 'name': CreateVM_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.340925] env[63531]: DEBUG oslo_vmware.api [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118043, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.086414} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.342537] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 875.343559] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d14fd92-ff0f-4241-902f-bb094ac0a584 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.365953] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Reconfiguring VM instance instance-00000048 to attach disk [datastore1] e95d3273-8216-47cc-95b6-99301366a827/e95d3273-8216-47cc-95b6-99301366a827.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 875.366981] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bb30d393-3189-48aa-a766-4c1006db552a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.386606] env[63531]: DEBUG oslo_vmware.api [None req-89c9bc13-a8c9-4b1a-9244-7179ff01bbf6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118034, 'name': CloneVM_Task, 'duration_secs': 1.284392} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.387248] env[63531]: INFO nova.virt.vmwareapi.vmops [None req-89c9bc13-a8c9-4b1a-9244-7179ff01bbf6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Created linked-clone VM from snapshot [ 875.388425] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51c138f7-8e08-4b9d-a0ab-783dc5ca624b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.393716] env[63531]: DEBUG oslo_vmware.api [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 875.393716] env[63531]: value = "task-1118045" [ 875.393716] env[63531]: _type = "Task" [ 875.393716] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.401298] env[63531]: DEBUG nova.virt.vmwareapi.images [None req-89c9bc13-a8c9-4b1a-9244-7179ff01bbf6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Uploading image f998b6ba-c103-40c9-a8fa-cecaa1f6d66c {{(pid=63531) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 875.409118] env[63531]: DEBUG oslo_vmware.api [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118045, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.433327] env[63531]: DEBUG oslo_vmware.rw_handles [None req-89c9bc13-a8c9-4b1a-9244-7179ff01bbf6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 875.433327] env[63531]: value = "vm-244727" [ 875.433327] env[63531]: _type = "VirtualMachine" [ 875.433327] env[63531]: }. {{(pid=63531) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 875.433732] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-fa995ca5-5cfa-4ded-b818-f8bf90509764 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.441594] env[63531]: DEBUG oslo_vmware.rw_handles [None req-89c9bc13-a8c9-4b1a-9244-7179ff01bbf6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lease: (returnval){ [ 875.441594] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]528548d0-4e16-8c62-ee2f-ef9baf6424cb" [ 875.441594] env[63531]: _type = "HttpNfcLease" [ 875.441594] env[63531]: } obtained for exporting VM: (result){ [ 875.441594] env[63531]: value = "vm-244727" [ 875.441594] env[63531]: _type = "VirtualMachine" [ 875.441594] env[63531]: }. {{(pid=63531) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 875.441931] env[63531]: DEBUG oslo_vmware.api [None req-89c9bc13-a8c9-4b1a-9244-7179ff01bbf6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the lease: (returnval){ [ 875.441931] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]528548d0-4e16-8c62-ee2f-ef9baf6424cb" [ 875.441931] env[63531]: _type = "HttpNfcLease" [ 875.441931] env[63531]: } to be ready. {{(pid=63531) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 875.451487] env[63531]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 875.451487] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]528548d0-4e16-8c62-ee2f-ef9baf6424cb" [ 875.451487] env[63531]: _type = "HttpNfcLease" [ 875.451487] env[63531]: } is initializing. {{(pid=63531) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 875.456389] env[63531]: DEBUG oslo_vmware.api [None req-545f958f-5555-4fdc-8d01-3ef53b1af173 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118044, 'name': ReconfigVM_Task, 'duration_secs': 0.278108} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.456688] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-545f958f-5555-4fdc-8d01-3ef53b1af173 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244725', 'volume_id': 'c89ce39a-7d3d-468e-b6d6-7df150ed8817', 'name': 'volume-c89ce39a-7d3d-468e-b6d6-7df150ed8817', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f46d8a85-6ab7-436a-bc5e-40d0f80974e5', 'attached_at': '', 'detached_at': '', 'volume_id': 'c89ce39a-7d3d-468e-b6d6-7df150ed8817', 'serial': 'c89ce39a-7d3d-468e-b6d6-7df150ed8817'} {{(pid=63531) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 875.588160] env[63531]: DEBUG oslo_vmware.api [None req-5ca07bd6-10c2-4158-87cf-a11b8952cc25 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]529d5c00-87c7-cc73-7f7d-b569cae2e282, 'name': SearchDatastore_Task, 'duration_secs': 0.011641} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.588160] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d7a5e783-d4dd-432c-af78-95e372cd6e76 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.593256] env[63531]: DEBUG oslo_vmware.api [None req-5ca07bd6-10c2-4158-87cf-a11b8952cc25 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 875.593256] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52326045-d5da-e6ff-5600-a8d107e10a80" [ 875.593256] env[63531]: _type = "Task" [ 875.593256] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.600767] env[63531]: DEBUG oslo_vmware.api [None req-5ca07bd6-10c2-4158-87cf-a11b8952cc25 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52326045-d5da-e6ff-5600-a8d107e10a80, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.627640] env[63531]: DEBUG nova.network.neutron [req-a7898dcb-d96d-4ec9-8d99-f5570e24c394 req-72aa5472-4e74-43ed-a6b5-dc8a9621095c service nova] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Updated VIF entry in instance network info cache for port 073adca7-1360-4703-9eb1-d86758bb8002. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 875.627976] env[63531]: DEBUG nova.network.neutron [req-a7898dcb-d96d-4ec9-8d99-f5570e24c394 req-72aa5472-4e74-43ed-a6b5-dc8a9621095c service nova] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Updating instance_info_cache with network_info: [{"id": "073adca7-1360-4703-9eb1-d86758bb8002", "address": "fa:16:3e:9b:a3:5c", "network": {"id": "95ed2c7d-3d1a-4736-8e90-104002f5c40b", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-73549344-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9056486e3b640d0bcdb9369ed5d67ad", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04e15990-16e1-4cb2-b0f0-06c362e68c5e", "external-id": "nsx-vlan-transportzone-555", "segmentation_id": 555, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap073adca7-13", "ovs_interfaceid": "073adca7-1360-4703-9eb1-d86758bb8002", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 875.735892] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118042, 'name': CreateVM_Task, 'duration_secs': 0.549629} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.736334] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 875.737222] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.737222] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.737482] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 875.737747] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-84aa5832-eba7-4db9-bdd3-91a1083dcae6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.745419] env[63531]: DEBUG oslo_vmware.api [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for the task: (returnval){ [ 875.745419] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52b32c21-4f5d-fe0e-480b-56ccf25eb958" [ 875.745419] env[63531]: _type = "Task" [ 875.745419] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.755468] env[63531]: DEBUG oslo_vmware.api [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52b32c21-4f5d-fe0e-480b-56ccf25eb958, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.792162] env[63531]: DEBUG nova.network.neutron [-] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 875.909293] env[63531]: DEBUG oslo_vmware.api [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118045, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.952999] env[63531]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 875.952999] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]528548d0-4e16-8c62-ee2f-ef9baf6424cb" [ 875.952999] env[63531]: _type = "HttpNfcLease" [ 875.952999] env[63531]: } is ready. {{(pid=63531) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 875.953844] env[63531]: DEBUG oslo_vmware.rw_handles [None req-89c9bc13-a8c9-4b1a-9244-7179ff01bbf6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 875.953844] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]528548d0-4e16-8c62-ee2f-ef9baf6424cb" [ 875.953844] env[63531]: _type = "HttpNfcLease" [ 875.953844] env[63531]: }. {{(pid=63531) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 875.955426] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da415759-34f2-4c45-aeab-a4656c6bb613 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.969841] env[63531]: DEBUG oslo_vmware.rw_handles [None req-89c9bc13-a8c9-4b1a-9244-7179ff01bbf6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/522b4d34-0b98-d713-f9e1-190e2d0581fd/disk-0.vmdk from lease info. {{(pid=63531) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 875.971170] env[63531]: DEBUG oslo_vmware.rw_handles [None req-89c9bc13-a8c9-4b1a-9244-7179ff01bbf6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/522b4d34-0b98-d713-f9e1-190e2d0581fd/disk-0.vmdk for reading. {{(pid=63531) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 876.068696] env[63531]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-7d45f288-4311-496b-8c6c-ad1a7ae6b471 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.081587] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30c6ee61-97e1-4f6a-a9eb-ffbf8e3f0515 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.090707] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acd72aad-bf1e-49ab-bb7c-8713839db989 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.102549] env[63531]: DEBUG nova.compute.manager [req-392ce068-0afe-4ff0-a1ff-69fb227f88cc req-ba6724a1-7046-4b4a-8d81-20d23dcbaf38 service nova] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Received event network-changed-2d27a8a5-c288-44fe-ac09-0d3b488386d6 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 876.102812] env[63531]: DEBUG nova.compute.manager [req-392ce068-0afe-4ff0-a1ff-69fb227f88cc req-ba6724a1-7046-4b4a-8d81-20d23dcbaf38 service nova] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Refreshing instance network info cache due to event network-changed-2d27a8a5-c288-44fe-ac09-0d3b488386d6. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 876.103116] env[63531]: DEBUG oslo_concurrency.lockutils [req-392ce068-0afe-4ff0-a1ff-69fb227f88cc req-ba6724a1-7046-4b4a-8d81-20d23dcbaf38 service nova] Acquiring lock "refresh_cache-2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.103308] env[63531]: DEBUG oslo_concurrency.lockutils [req-392ce068-0afe-4ff0-a1ff-69fb227f88cc req-ba6724a1-7046-4b4a-8d81-20d23dcbaf38 service nova] Acquired lock "refresh_cache-2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.103516] env[63531]: DEBUG nova.network.neutron [req-392ce068-0afe-4ff0-a1ff-69fb227f88cc req-ba6724a1-7046-4b4a-8d81-20d23dcbaf38 service nova] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Refreshing network info cache for port 2d27a8a5-c288-44fe-ac09-0d3b488386d6 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 876.132019] env[63531]: DEBUG oslo_concurrency.lockutils [req-a7898dcb-d96d-4ec9-8d99-f5570e24c394 req-72aa5472-4e74-43ed-a6b5-dc8a9621095c service nova] Releasing lock "refresh_cache-37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.132413] env[63531]: DEBUG nova.compute.manager [req-a7898dcb-d96d-4ec9-8d99-f5570e24c394 req-72aa5472-4e74-43ed-a6b5-dc8a9621095c service nova] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Received event network-vif-unplugged-2d27a8a5-c288-44fe-ac09-0d3b488386d6 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 876.133123] env[63531]: DEBUG oslo_concurrency.lockutils [req-a7898dcb-d96d-4ec9-8d99-f5570e24c394 req-72aa5472-4e74-43ed-a6b5-dc8a9621095c service nova] Acquiring lock "2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.133413] env[63531]: DEBUG oslo_concurrency.lockutils [req-a7898dcb-d96d-4ec9-8d99-f5570e24c394 req-72aa5472-4e74-43ed-a6b5-dc8a9621095c service nova] Lock "2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.134034] env[63531]: DEBUG oslo_concurrency.lockutils [req-a7898dcb-d96d-4ec9-8d99-f5570e24c394 req-72aa5472-4e74-43ed-a6b5-dc8a9621095c service nova] Lock "2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.134034] env[63531]: DEBUG nova.compute.manager [req-a7898dcb-d96d-4ec9-8d99-f5570e24c394 req-72aa5472-4e74-43ed-a6b5-dc8a9621095c service nova] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] No waiting events found dispatching network-vif-unplugged-2d27a8a5-c288-44fe-ac09-0d3b488386d6 {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 876.134034] env[63531]: WARNING nova.compute.manager [req-a7898dcb-d96d-4ec9-8d99-f5570e24c394 req-72aa5472-4e74-43ed-a6b5-dc8a9621095c service nova] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Received unexpected event network-vif-unplugged-2d27a8a5-c288-44fe-ac09-0d3b488386d6 for instance with vm_state shelved and task_state shelving_offloading. [ 876.141753] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f37d87f4-13fd-41cd-9266-ba72f031c4b2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.144900] env[63531]: DEBUG oslo_vmware.api [None req-5ca07bd6-10c2-4158-87cf-a11b8952cc25 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52326045-d5da-e6ff-5600-a8d107e10a80, 'name': SearchDatastore_Task, 'duration_secs': 0.009623} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.145444] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5ca07bd6-10c2-4158-87cf-a11b8952cc25 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.146010] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ca07bd6-10c2-4158-87cf-a11b8952cc25 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] 879924d4-b465-4102-a0e3-c7b2be7ef08b/9d550399-588c-4bd5-8436-eba9b523bd13-rescue.vmdk. {{(pid=63531) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 876.147163] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a6acd213-b970-4562-ae95-0208c25d89a0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.159550] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24a073fe-b5c9-4513-a63e-7a89f646acfc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.166724] env[63531]: DEBUG oslo_vmware.api [None req-5ca07bd6-10c2-4158-87cf-a11b8952cc25 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 876.166724] env[63531]: value = "task-1118047" [ 876.166724] env[63531]: _type = "Task" [ 876.166724] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.179281] env[63531]: DEBUG nova.compute.provider_tree [None req-1b340bf7-08de-4d21-a861-c6e587976b10 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 876.186599] env[63531]: DEBUG oslo_vmware.api [None req-5ca07bd6-10c2-4158-87cf-a11b8952cc25 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118047, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.262527] env[63531]: DEBUG oslo_vmware.api [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52b32c21-4f5d-fe0e-480b-56ccf25eb958, 'name': SearchDatastore_Task, 'duration_secs': 0.021044} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.262527] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.262527] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 876.262527] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.262527] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.262527] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 876.262527] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2415d945-2cd7-4a2a-ab7e-a5d4d24cbaf1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.273511] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 876.273892] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 876.274638] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6b8ad1b3-0993-4491-88e1-9e51ccc9c3fb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.283973] env[63531]: DEBUG oslo_vmware.api [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for the task: (returnval){ [ 876.283973] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5298b69f-c97e-3a11-bc66-0fac31606db5" [ 876.283973] env[63531]: _type = "Task" [ 876.283973] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.294523] env[63531]: INFO nova.compute.manager [-] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Took 1.28 seconds to deallocate network for instance. [ 876.294902] env[63531]: DEBUG oslo_vmware.api [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5298b69f-c97e-3a11-bc66-0fac31606db5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.406253] env[63531]: DEBUG oslo_vmware.api [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118045, 'name': ReconfigVM_Task, 'duration_secs': 0.704314} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.406572] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Reconfigured VM instance instance-00000048 to attach disk [datastore1] e95d3273-8216-47cc-95b6-99301366a827/e95d3273-8216-47cc-95b6-99301366a827.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 876.406930] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4daf3e9b-0dfd-47d2-8169-011a57237f05 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.416036] env[63531]: DEBUG oslo_vmware.api [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 876.416036] env[63531]: value = "task-1118048" [ 876.416036] env[63531]: _type = "Task" [ 876.416036] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.427280] env[63531]: DEBUG oslo_vmware.api [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118048, 'name': Rename_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.537280] env[63531]: DEBUG nova.objects.instance [None req-545f958f-5555-4fdc-8d01-3ef53b1af173 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lazy-loading 'flavor' on Instance uuid f46d8a85-6ab7-436a-bc5e-40d0f80974e5 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 876.682618] env[63531]: DEBUG nova.scheduler.client.report [None req-1b340bf7-08de-4d21-a861-c6e587976b10 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 876.701852] env[63531]: DEBUG oslo_vmware.api [None req-5ca07bd6-10c2-4158-87cf-a11b8952cc25 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118047, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.803499] env[63531]: DEBUG oslo_vmware.api [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5298b69f-c97e-3a11-bc66-0fac31606db5, 'name': SearchDatastore_Task, 'duration_secs': 0.019433} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.804610] env[63531]: DEBUG oslo_concurrency.lockutils [None req-29b04bcd-cd4b-4a53-9b9f-4a106df45c8e tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.805654] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a23f02af-98bb-404e-897c-9fe96cb313b5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.817185] env[63531]: DEBUG oslo_vmware.api [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for the task: (returnval){ [ 876.817185] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52072ef4-428c-e754-6127-ebc50486dbc0" [ 876.817185] env[63531]: _type = "Task" [ 876.817185] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.828757] env[63531]: DEBUG oslo_vmware.api [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52072ef4-428c-e754-6127-ebc50486dbc0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.931939] env[63531]: DEBUG oslo_vmware.api [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118048, 'name': Rename_Task, 'duration_secs': 0.341931} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.932810] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 876.934047] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c3ab4e8f-6719-4070-83cb-db27be0827b9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.944827] env[63531]: DEBUG oslo_vmware.api [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 876.944827] env[63531]: value = "task-1118049" [ 876.944827] env[63531]: _type = "Task" [ 876.944827] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.958134] env[63531]: DEBUG oslo_vmware.api [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118049, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.043858] env[63531]: DEBUG oslo_concurrency.lockutils [None req-545f958f-5555-4fdc-8d01-3ef53b1af173 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lock "f46d8a85-6ab7-436a-bc5e-40d0f80974e5" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.376s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.130903] env[63531]: DEBUG nova.network.neutron [req-392ce068-0afe-4ff0-a1ff-69fb227f88cc req-ba6724a1-7046-4b4a-8d81-20d23dcbaf38 service nova] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Updated VIF entry in instance network info cache for port 2d27a8a5-c288-44fe-ac09-0d3b488386d6. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 877.131329] env[63531]: DEBUG nova.network.neutron [req-392ce068-0afe-4ff0-a1ff-69fb227f88cc req-ba6724a1-7046-4b4a-8d81-20d23dcbaf38 service nova] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Updating instance_info_cache with network_info: [{"id": "2d27a8a5-c288-44fe-ac09-0d3b488386d6", "address": "fa:16:3e:91:1c:62", "network": {"id": "5b8b7134-f123-40b8-9134-5cd072e8ad01", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1232565840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.165", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef3b56ce3f3a4155bd09642f5c48ea6f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap2d27a8a5-c2", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 877.181937] env[63531]: DEBUG oslo_vmware.api [None req-5ca07bd6-10c2-4158-87cf-a11b8952cc25 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118047, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.65071} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.182290] env[63531]: INFO nova.virt.vmwareapi.ds_util [None req-5ca07bd6-10c2-4158-87cf-a11b8952cc25 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] 879924d4-b465-4102-a0e3-c7b2be7ef08b/9d550399-588c-4bd5-8436-eba9b523bd13-rescue.vmdk. [ 877.183250] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08395c09-3891-4ca6-8555-ad661e7eb6c7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.190378] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1b340bf7-08de-4d21-a861-c6e587976b10 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.042s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.208397] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 10.283s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.208689] env[63531]: DEBUG nova.objects.instance [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63531) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 877.219844] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ca07bd6-10c2-4158-87cf-a11b8952cc25 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Reconfiguring VM instance instance-00000047 to attach disk [datastore2] 879924d4-b465-4102-a0e3-c7b2be7ef08b/9d550399-588c-4bd5-8436-eba9b523bd13-rescue.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 877.220606] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f8521970-203b-42e3-95d8-1d67de3c0ad4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.238107] env[63531]: INFO nova.scheduler.client.report [None req-1b340bf7-08de-4d21-a861-c6e587976b10 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Deleted allocations for instance 09feb12a-da9f-4bb7-959d-84c5c7c58e99 [ 877.254815] env[63531]: DEBUG oslo_vmware.api [None req-5ca07bd6-10c2-4158-87cf-a11b8952cc25 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 877.254815] env[63531]: value = "task-1118050" [ 877.254815] env[63531]: _type = "Task" [ 877.254815] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.268044] env[63531]: DEBUG oslo_vmware.api [None req-5ca07bd6-10c2-4158-87cf-a11b8952cc25 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118050, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.329327] env[63531]: DEBUG oslo_vmware.api [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52072ef4-428c-e754-6127-ebc50486dbc0, 'name': SearchDatastore_Task, 'duration_secs': 0.016598} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.330558] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.330558] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9/37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 877.330558] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2b1daec8-fe5d-4448-9dad-cf204975c6a1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.342731] env[63531]: DEBUG oslo_vmware.api [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for the task: (returnval){ [ 877.342731] env[63531]: value = "task-1118051" [ 877.342731] env[63531]: _type = "Task" [ 877.342731] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.353845] env[63531]: DEBUG oslo_vmware.api [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118051, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.459210] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9c5e2adc-293d-4872-833f-c9ef19698f71 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Acquiring lock "f46d8a85-6ab7-436a-bc5e-40d0f80974e5" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.459621] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9c5e2adc-293d-4872-833f-c9ef19698f71 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lock "f46d8a85-6ab7-436a-bc5e-40d0f80974e5" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.459858] env[63531]: DEBUG nova.compute.manager [None req-9c5e2adc-293d-4872-833f-c9ef19698f71 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 877.460258] env[63531]: DEBUG oslo_vmware.api [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118049, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.461281] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d44fd21-d866-437f-bf9a-1f7cb8ca950c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.469574] env[63531]: DEBUG nova.compute.manager [None req-9c5e2adc-293d-4872-833f-c9ef19698f71 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63531) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 877.470694] env[63531]: DEBUG nova.objects.instance [None req-9c5e2adc-293d-4872-833f-c9ef19698f71 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lazy-loading 'flavor' on Instance uuid f46d8a85-6ab7-436a-bc5e-40d0f80974e5 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 877.569753] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Acquiring lock "2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.634461] env[63531]: DEBUG oslo_concurrency.lockutils [req-392ce068-0afe-4ff0-a1ff-69fb227f88cc req-ba6724a1-7046-4b4a-8d81-20d23dcbaf38 service nova] Releasing lock "refresh_cache-2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.635041] env[63531]: DEBUG nova.compute.manager [req-392ce068-0afe-4ff0-a1ff-69fb227f88cc req-ba6724a1-7046-4b4a-8d81-20d23dcbaf38 service nova] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Received event network-vif-deleted-f02115dd-c7b4-46f6-8057-20f0e4985b2e {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 877.749767] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1b340bf7-08de-4d21-a861-c6e587976b10 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Lock "09feb12a-da9f-4bb7-959d-84c5c7c58e99" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.120s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.769440] env[63531]: DEBUG oslo_vmware.api [None req-5ca07bd6-10c2-4158-87cf-a11b8952cc25 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118050, 'name': ReconfigVM_Task, 'duration_secs': 0.39284} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.770637] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ca07bd6-10c2-4158-87cf-a11b8952cc25 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Reconfigured VM instance instance-00000047 to attach disk [datastore2] 879924d4-b465-4102-a0e3-c7b2be7ef08b/9d550399-588c-4bd5-8436-eba9b523bd13-rescue.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 877.771320] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6b391a2-bc55-4461-b55c-cd5e5696bfa8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.801087] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d644b60d-f8fa-4a0b-8655-016d01577bd7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.819722] env[63531]: DEBUG oslo_vmware.api [None req-5ca07bd6-10c2-4158-87cf-a11b8952cc25 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 877.819722] env[63531]: value = "task-1118052" [ 877.819722] env[63531]: _type = "Task" [ 877.819722] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.831387] env[63531]: DEBUG oslo_vmware.api [None req-5ca07bd6-10c2-4158-87cf-a11b8952cc25 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118052, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.858577] env[63531]: DEBUG oslo_vmware.api [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118051, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.959546] env[63531]: DEBUG oslo_vmware.api [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118049, 'name': PowerOnVM_Task, 'duration_secs': 0.567587} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.960109] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 877.960459] env[63531]: INFO nova.compute.manager [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Took 12.99 seconds to spawn the instance on the hypervisor. [ 877.960684] env[63531]: DEBUG nova.compute.manager [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 877.961604] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0771a8b7-623a-475a-adfc-431bd1882df2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.978770] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c5e2adc-293d-4872-833f-c9ef19698f71 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 877.980240] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-30d25222-7089-4f5d-8427-9c33ed8b0982 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.988462] env[63531]: DEBUG oslo_vmware.api [None req-9c5e2adc-293d-4872-833f-c9ef19698f71 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Waiting for the task: (returnval){ [ 877.988462] env[63531]: value = "task-1118053" [ 877.988462] env[63531]: _type = "Task" [ 877.988462] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.000648] env[63531]: DEBUG oslo_vmware.api [None req-9c5e2adc-293d-4872-833f-c9ef19698f71 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118053, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.218881] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5cb5a392-5b9e-4716-90c7-5bd2065ea2ad tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.010s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.222324] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f2e31732-dbaf-4d48-baef-afe717b5167a tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.432s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.222324] env[63531]: DEBUG nova.objects.instance [None req-f2e31732-dbaf-4d48-baef-afe717b5167a tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Lazy-loading 'resources' on Instance uuid 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 878.332102] env[63531]: DEBUG oslo_vmware.api [None req-5ca07bd6-10c2-4158-87cf-a11b8952cc25 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118052, 'name': ReconfigVM_Task, 'duration_secs': 0.254257} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.335242] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ca07bd6-10c2-4158-87cf-a11b8952cc25 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 878.335242] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-47877e49-ef8f-415b-884e-96c4567ba12b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.341719] env[63531]: DEBUG oslo_vmware.api [None req-5ca07bd6-10c2-4158-87cf-a11b8952cc25 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 878.341719] env[63531]: value = "task-1118054" [ 878.341719] env[63531]: _type = "Task" [ 878.341719] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.354978] env[63531]: DEBUG oslo_vmware.api [None req-5ca07bd6-10c2-4158-87cf-a11b8952cc25 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118054, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.359053] env[63531]: DEBUG oslo_vmware.api [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118051, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.656853} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.359475] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9/37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 878.359725] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 878.360044] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fef168b8-b1f1-489d-b4eb-bb836cbcf629 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.368430] env[63531]: DEBUG oslo_vmware.api [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for the task: (returnval){ [ 878.368430] env[63531]: value = "task-1118055" [ 878.368430] env[63531]: _type = "Task" [ 878.368430] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.378755] env[63531]: DEBUG oslo_vmware.api [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118055, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.399559] env[63531]: DEBUG oslo_concurrency.lockutils [None req-31ea6a4d-757c-44c3-8294-8cbb5267d005 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Acquiring lock "3eeaa6f9-60d9-43fe-a438-fbf2ceececae" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.399951] env[63531]: DEBUG oslo_concurrency.lockutils [None req-31ea6a4d-757c-44c3-8294-8cbb5267d005 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Lock "3eeaa6f9-60d9-43fe-a438-fbf2ceececae" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.400199] env[63531]: DEBUG oslo_concurrency.lockutils [None req-31ea6a4d-757c-44c3-8294-8cbb5267d005 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Acquiring lock "3eeaa6f9-60d9-43fe-a438-fbf2ceececae-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.400414] env[63531]: DEBUG oslo_concurrency.lockutils [None req-31ea6a4d-757c-44c3-8294-8cbb5267d005 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Lock "3eeaa6f9-60d9-43fe-a438-fbf2ceececae-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.400580] env[63531]: DEBUG oslo_concurrency.lockutils [None req-31ea6a4d-757c-44c3-8294-8cbb5267d005 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Lock "3eeaa6f9-60d9-43fe-a438-fbf2ceececae-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.402898] env[63531]: INFO nova.compute.manager [None req-31ea6a4d-757c-44c3-8294-8cbb5267d005 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 3eeaa6f9-60d9-43fe-a438-fbf2ceececae] Terminating instance [ 878.406770] env[63531]: DEBUG oslo_concurrency.lockutils [None req-31ea6a4d-757c-44c3-8294-8cbb5267d005 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Acquiring lock "refresh_cache-3eeaa6f9-60d9-43fe-a438-fbf2ceececae" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.406866] env[63531]: DEBUG oslo_concurrency.lockutils [None req-31ea6a4d-757c-44c3-8294-8cbb5267d005 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Acquired lock "refresh_cache-3eeaa6f9-60d9-43fe-a438-fbf2ceececae" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.407055] env[63531]: DEBUG nova.network.neutron [None req-31ea6a4d-757c-44c3-8294-8cbb5267d005 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 3eeaa6f9-60d9-43fe-a438-fbf2ceececae] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 878.484454] env[63531]: INFO nova.compute.manager [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Took 41.70 seconds to build instance. [ 878.501486] env[63531]: DEBUG oslo_vmware.api [None req-9c5e2adc-293d-4872-833f-c9ef19698f71 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118053, 'name': PowerOffVM_Task, 'duration_secs': 0.264568} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.502176] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c5e2adc-293d-4872-833f-c9ef19698f71 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 878.502382] env[63531]: DEBUG nova.compute.manager [None req-9c5e2adc-293d-4872-833f-c9ef19698f71 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 878.503418] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec99fe5f-59bc-4ee2-a680-68e73b6eca09 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.855009] env[63531]: DEBUG oslo_vmware.api [None req-5ca07bd6-10c2-4158-87cf-a11b8952cc25 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118054, 'name': PowerOnVM_Task, 'duration_secs': 0.512633} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.858696] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ca07bd6-10c2-4158-87cf-a11b8952cc25 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 878.862545] env[63531]: DEBUG nova.compute.manager [None req-5ca07bd6-10c2-4158-87cf-a11b8952cc25 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 878.863843] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49c2174d-7885-4db3-b0b9-00a0151cd42e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.891244] env[63531]: DEBUG oslo_vmware.api [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118055, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.108492} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.891856] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 878.893275] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56e313a5-374e-49b1-8aa1-5669a4a85e9a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.931792] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Reconfiguring VM instance instance-00000049 to attach disk [datastore1] 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9/37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 878.937251] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c15975e8-d070-40f7-9e3f-dd05b546abdc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.963934] env[63531]: DEBUG oslo_vmware.api [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for the task: (returnval){ [ 878.963934] env[63531]: value = "task-1118056" [ 878.963934] env[63531]: _type = "Task" [ 878.963934] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.976604] env[63531]: DEBUG oslo_vmware.api [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118056, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.986457] env[63531]: DEBUG oslo_concurrency.lockutils [None req-29843cc2-9211-4f0c-9f78-5af704d87eb1 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "e95d3273-8216-47cc-95b6-99301366a827" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 59.217s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.991200] env[63531]: DEBUG nova.network.neutron [None req-31ea6a4d-757c-44c3-8294-8cbb5267d005 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 3eeaa6f9-60d9-43fe-a438-fbf2ceececae] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 879.022343] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9c5e2adc-293d-4872-833f-c9ef19698f71 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lock "f46d8a85-6ab7-436a-bc5e-40d0f80974e5" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.563s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.104281] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b835249-6e26-418a-b857-4d7227cc85c6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.113267] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c94f551e-4b8b-462a-8f0b-74fd3e5948b4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.119179] env[63531]: DEBUG nova.network.neutron [None req-31ea6a4d-757c-44c3-8294-8cbb5267d005 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 3eeaa6f9-60d9-43fe-a438-fbf2ceececae] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 879.149712] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c3d25e5-4366-41f7-bc50-8929be64d57e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.159187] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9347e46c-02a9-48ae-aebf-366775d47046 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.180835] env[63531]: DEBUG nova.compute.provider_tree [None req-f2e31732-dbaf-4d48-baef-afe717b5167a tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 879.475323] env[63531]: DEBUG oslo_vmware.api [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118056, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.622996] env[63531]: DEBUG oslo_concurrency.lockutils [None req-31ea6a4d-757c-44c3-8294-8cbb5267d005 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Releasing lock "refresh_cache-3eeaa6f9-60d9-43fe-a438-fbf2ceececae" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.623494] env[63531]: DEBUG nova.compute.manager [None req-31ea6a4d-757c-44c3-8294-8cbb5267d005 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 3eeaa6f9-60d9-43fe-a438-fbf2ceececae] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 879.623699] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-31ea6a4d-757c-44c3-8294-8cbb5267d005 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 3eeaa6f9-60d9-43fe-a438-fbf2ceececae] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 879.624593] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5db27ecf-9788-4ff5-8b9a-5a751a2d6700 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.633867] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-31ea6a4d-757c-44c3-8294-8cbb5267d005 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 3eeaa6f9-60d9-43fe-a438-fbf2ceececae] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 879.634133] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-47a723f7-e79b-4e1f-8aab-c62b98ed7ac9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.641958] env[63531]: DEBUG oslo_vmware.api [None req-31ea6a4d-757c-44c3-8294-8cbb5267d005 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Waiting for the task: (returnval){ [ 879.641958] env[63531]: value = "task-1118057" [ 879.641958] env[63531]: _type = "Task" [ 879.641958] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.651240] env[63531]: DEBUG oslo_vmware.api [None req-31ea6a4d-757c-44c3-8294-8cbb5267d005 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': task-1118057, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.688028] env[63531]: DEBUG nova.scheduler.client.report [None req-f2e31732-dbaf-4d48-baef-afe717b5167a tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 879.820038] env[63531]: DEBUG oslo_concurrency.lockutils [None req-cb9d9d6a-7e5e-4565-b903-98da7cb60994 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquiring lock "e95d3273-8216-47cc-95b6-99301366a827" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.820038] env[63531]: DEBUG oslo_concurrency.lockutils [None req-cb9d9d6a-7e5e-4565-b903-98da7cb60994 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "e95d3273-8216-47cc-95b6-99301366a827" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.820231] env[63531]: DEBUG nova.compute.manager [None req-cb9d9d6a-7e5e-4565-b903-98da7cb60994 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 879.821172] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d6271d0-6045-4d7b-995f-ba9365639168 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.828546] env[63531]: DEBUG nova.compute.manager [None req-cb9d9d6a-7e5e-4565-b903-98da7cb60994 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63531) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 879.829203] env[63531]: DEBUG nova.objects.instance [None req-cb9d9d6a-7e5e-4565-b903-98da7cb60994 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lazy-loading 'flavor' on Instance uuid e95d3273-8216-47cc-95b6-99301366a827 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 879.976979] env[63531]: DEBUG oslo_vmware.api [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118056, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.990111] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Acquiring lock "1ca7b851-2a06-4181-8271-58aafcd322d6" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.990396] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Lock "1ca7b851-2a06-4181-8271-58aafcd322d6" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.991083] env[63531]: INFO nova.compute.manager [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Shelving [ 880.155453] env[63531]: DEBUG oslo_vmware.api [None req-31ea6a4d-757c-44c3-8294-8cbb5267d005 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': task-1118057, 'name': PowerOffVM_Task, 'duration_secs': 0.174997} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.155806] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-31ea6a4d-757c-44c3-8294-8cbb5267d005 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 3eeaa6f9-60d9-43fe-a438-fbf2ceececae] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 880.156037] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-31ea6a4d-757c-44c3-8294-8cbb5267d005 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 3eeaa6f9-60d9-43fe-a438-fbf2ceececae] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 880.156350] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5b5af7eb-457e-4ed1-bad4-20581baf26be {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.191592] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-31ea6a4d-757c-44c3-8294-8cbb5267d005 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 3eeaa6f9-60d9-43fe-a438-fbf2ceececae] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 880.191976] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-31ea6a4d-757c-44c3-8294-8cbb5267d005 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 3eeaa6f9-60d9-43fe-a438-fbf2ceececae] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 880.192339] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-31ea6a4d-757c-44c3-8294-8cbb5267d005 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Deleting the datastore file [datastore1] 3eeaa6f9-60d9-43fe-a438-fbf2ceececae {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 880.193504] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f2e31732-dbaf-4d48-baef-afe717b5167a tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.973s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.196834] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1016b05c-da0d-4526-9fe4-b651913d5ca7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.200563] env[63531]: DEBUG oslo_concurrency.lockutils [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.924s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.203170] env[63531]: INFO nova.compute.claims [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 880.215222] env[63531]: DEBUG oslo_vmware.api [None req-31ea6a4d-757c-44c3-8294-8cbb5267d005 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Waiting for the task: (returnval){ [ 880.215222] env[63531]: value = "task-1118059" [ 880.215222] env[63531]: _type = "Task" [ 880.215222] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.224962] env[63531]: DEBUG oslo_vmware.api [None req-31ea6a4d-757c-44c3-8294-8cbb5267d005 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': task-1118059, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.242409] env[63531]: INFO nova.scheduler.client.report [None req-f2e31732-dbaf-4d48-baef-afe717b5167a tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Deleted allocations for instance 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b [ 880.338048] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb9d9d6a-7e5e-4565-b903-98da7cb60994 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 880.338369] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-edba2693-4082-472e-8dd1-35e051b22ede {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.348138] env[63531]: DEBUG oslo_vmware.api [None req-cb9d9d6a-7e5e-4565-b903-98da7cb60994 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 880.348138] env[63531]: value = "task-1118060" [ 880.348138] env[63531]: _type = "Task" [ 880.348138] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.358119] env[63531]: DEBUG oslo_vmware.api [None req-cb9d9d6a-7e5e-4565-b903-98da7cb60994 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118060, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.476136] env[63531]: DEBUG oslo_vmware.api [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118056, 'name': ReconfigVM_Task, 'duration_secs': 1.128793} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.476505] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Reconfigured VM instance instance-00000049 to attach disk [datastore1] 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9/37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 880.477185] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6380abd4-2d88-4ec8-abca-bdc7b75cb086 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.486285] env[63531]: DEBUG oslo_vmware.api [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for the task: (returnval){ [ 880.486285] env[63531]: value = "task-1118061" [ 880.486285] env[63531]: _type = "Task" [ 880.486285] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.500397] env[63531]: DEBUG oslo_vmware.api [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118061, 'name': Rename_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.504375] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 880.504375] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-47c7640f-be7a-4933-ab2e-91216e6f2196 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.511842] env[63531]: DEBUG oslo_vmware.api [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Waiting for the task: (returnval){ [ 880.511842] env[63531]: value = "task-1118062" [ 880.511842] env[63531]: _type = "Task" [ 880.511842] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.522459] env[63531]: DEBUG oslo_vmware.api [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1118062, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.589458] env[63531]: DEBUG nova.objects.instance [None req-0365d10b-233e-47b3-a10a-1ec120c16583 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lazy-loading 'flavor' on Instance uuid f46d8a85-6ab7-436a-bc5e-40d0f80974e5 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 880.727504] env[63531]: DEBUG oslo_vmware.api [None req-31ea6a4d-757c-44c3-8294-8cbb5267d005 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Task: {'id': task-1118059, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.14728} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.727779] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-31ea6a4d-757c-44c3-8294-8cbb5267d005 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 880.727975] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-31ea6a4d-757c-44c3-8294-8cbb5267d005 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 3eeaa6f9-60d9-43fe-a438-fbf2ceececae] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 880.728176] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-31ea6a4d-757c-44c3-8294-8cbb5267d005 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 3eeaa6f9-60d9-43fe-a438-fbf2ceececae] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 880.728373] env[63531]: INFO nova.compute.manager [None req-31ea6a4d-757c-44c3-8294-8cbb5267d005 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] [instance: 3eeaa6f9-60d9-43fe-a438-fbf2ceececae] Took 1.10 seconds to destroy the instance on the hypervisor. [ 880.728647] env[63531]: DEBUG oslo.service.loopingcall [None req-31ea6a4d-757c-44c3-8294-8cbb5267d005 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 880.728849] env[63531]: DEBUG nova.compute.manager [-] [instance: 3eeaa6f9-60d9-43fe-a438-fbf2ceececae] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 880.728944] env[63531]: DEBUG nova.network.neutron [-] [instance: 3eeaa6f9-60d9-43fe-a438-fbf2ceececae] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 880.748570] env[63531]: DEBUG nova.network.neutron [-] [instance: 3eeaa6f9-60d9-43fe-a438-fbf2ceececae] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 880.754040] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f2e31732-dbaf-4d48-baef-afe717b5167a tempest-ServersTestMultiNic-569120898 tempest-ServersTestMultiNic-569120898-project-member] Lock "92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.659s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.861013] env[63531]: DEBUG oslo_vmware.api [None req-cb9d9d6a-7e5e-4565-b903-98da7cb60994 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118060, 'name': PowerOffVM_Task, 'duration_secs': 0.213645} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.861424] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb9d9d6a-7e5e-4565-b903-98da7cb60994 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 880.861727] env[63531]: DEBUG nova.compute.manager [None req-cb9d9d6a-7e5e-4565-b903-98da7cb60994 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 880.862700] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46ab701b-d67c-4552-8b1b-f3a3a0539577 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.997658] env[63531]: DEBUG oslo_vmware.api [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118061, 'name': Rename_Task, 'duration_secs': 0.268731} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.997985] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 880.998293] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5853bcf3-1aa6-4d8a-ac2c-cc0d8b92044b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.007166] env[63531]: DEBUG oslo_vmware.api [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for the task: (returnval){ [ 881.007166] env[63531]: value = "task-1118063" [ 881.007166] env[63531]: _type = "Task" [ 881.007166] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.019768] env[63531]: DEBUG oslo_vmware.api [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118063, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.026373] env[63531]: DEBUG oslo_vmware.api [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1118062, 'name': PowerOffVM_Task, 'duration_secs': 0.427009} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.026875] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 881.027718] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b708da50-ac52-4a99-8a79-b5dddab51e00 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.050255] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2da8e618-4296-4c16-b2f7-27360ae4e6db {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.094978] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0365d10b-233e-47b3-a10a-1ec120c16583 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Acquiring lock "refresh_cache-f46d8a85-6ab7-436a-bc5e-40d0f80974e5" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.095716] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0365d10b-233e-47b3-a10a-1ec120c16583 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Acquired lock "refresh_cache-f46d8a85-6ab7-436a-bc5e-40d0f80974e5" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.095716] env[63531]: DEBUG nova.network.neutron [None req-0365d10b-233e-47b3-a10a-1ec120c16583 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 881.095863] env[63531]: DEBUG nova.objects.instance [None req-0365d10b-233e-47b3-a10a-1ec120c16583 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lazy-loading 'info_cache' on Instance uuid f46d8a85-6ab7-436a-bc5e-40d0f80974e5 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 881.253537] env[63531]: DEBUG nova.network.neutron [-] [instance: 3eeaa6f9-60d9-43fe-a438-fbf2ceececae] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.376609] env[63531]: DEBUG oslo_concurrency.lockutils [None req-cb9d9d6a-7e5e-4565-b903-98da7cb60994 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "e95d3273-8216-47cc-95b6-99301366a827" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.557s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.518105] env[63531]: DEBUG oslo_vmware.api [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118063, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.519789] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c511b3cd-0eb5-47c9-9d14-0a9b66ec15a6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.528037] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16a3e2c6-1695-463d-91ee-3a1e39249a25 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.563832] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Creating Snapshot of the VM instance {{(pid=63531) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 881.564440] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-b0a23814-7d46-45b0-bcb8-11f2f520baa3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.567512] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e593de6b-97e9-4145-b0df-c416a5ef8555 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.581556] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35b12928-9ea8-4c4d-81db-bacfadcbdab1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.586480] env[63531]: DEBUG oslo_vmware.api [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Waiting for the task: (returnval){ [ 881.586480] env[63531]: value = "task-1118064" [ 881.586480] env[63531]: _type = "Task" [ 881.586480] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.602266] env[63531]: DEBUG nova.objects.base [None req-0365d10b-233e-47b3-a10a-1ec120c16583 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=63531) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 881.604252] env[63531]: DEBUG nova.compute.provider_tree [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 881.612206] env[63531]: DEBUG oslo_vmware.api [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1118064, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.692081] env[63531]: INFO nova.compute.manager [None req-7e54157c-b2af-4c09-ad30-d81f45506a37 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Rescuing [ 881.692551] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7e54157c-b2af-4c09-ad30-d81f45506a37 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquiring lock "refresh_cache-4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.692743] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7e54157c-b2af-4c09-ad30-d81f45506a37 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquired lock "refresh_cache-4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.692924] env[63531]: DEBUG nova.network.neutron [None req-7e54157c-b2af-4c09-ad30-d81f45506a37 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 881.756753] env[63531]: INFO nova.compute.manager [-] [instance: 3eeaa6f9-60d9-43fe-a438-fbf2ceececae] Took 1.03 seconds to deallocate network for instance. [ 882.019604] env[63531]: DEBUG oslo_vmware.api [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118063, 'name': PowerOnVM_Task, 'duration_secs': 0.654506} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.019831] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 882.020035] env[63531]: INFO nova.compute.manager [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Took 9.89 seconds to spawn the instance on the hypervisor. [ 882.020287] env[63531]: DEBUG nova.compute.manager [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 882.021092] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6186a72e-2763-49a2-a4ca-e4c4189a023c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.097506] env[63531]: DEBUG oslo_vmware.api [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1118064, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.109669] env[63531]: DEBUG nova.scheduler.client.report [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 882.265349] env[63531]: DEBUG oslo_concurrency.lockutils [None req-31ea6a4d-757c-44c3-8294-8cbb5267d005 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.436123] env[63531]: DEBUG nova.network.neutron [None req-7e54157c-b2af-4c09-ad30-d81f45506a37 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Updating instance_info_cache with network_info: [{"id": "e30711c3-e609-434b-82da-a58252d3678d", "address": "fa:16:3e:54:c1:b7", "network": {"id": "913045d9-e90b-4366-987c-7706b73a6147", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1696510389-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f71c7db0a3614995becc25d9a8dce787", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c297fe21-cd0b-4226-813b-a65d2358d034", "external-id": "nsx-vlan-transportzone-98", "segmentation_id": 98, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape30711c3-e6", "ovs_interfaceid": "e30711c3-e609-434b-82da-a58252d3678d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 882.484157] env[63531]: DEBUG nova.network.neutron [None req-0365d10b-233e-47b3-a10a-1ec120c16583 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Updating instance_info_cache with network_info: [{"id": "a3fd640b-c47b-4501-ac04-f90f6af8b076", "address": "fa:16:3e:27:ca:51", "network": {"id": "4dd26082-3411-4d99-a78b-c76c15703506", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1190890625-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.196", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ee5fd84e1da04fd984f81578e4085850", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6046aec4-feda-4ef9-bf4a-800de1e0cd3b", "external-id": "nsx-vlan-transportzone-903", "segmentation_id": 903, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3fd640b-c4", "ovs_interfaceid": "a3fd640b-c47b-4501-ac04-f90f6af8b076", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 882.540327] env[63531]: INFO nova.compute.manager [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Took 33.67 seconds to build instance. [ 882.598036] env[63531]: DEBUG oslo_vmware.api [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1118064, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.615721] env[63531]: DEBUG oslo_concurrency.lockutils [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.415s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.616426] env[63531]: DEBUG nova.compute.manager [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 882.619588] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.358s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.621173] env[63531]: INFO nova.compute.claims [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 882.940730] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7e54157c-b2af-4c09-ad30-d81f45506a37 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Releasing lock "refresh_cache-4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.987047] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0365d10b-233e-47b3-a10a-1ec120c16583 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Releasing lock "refresh_cache-f46d8a85-6ab7-436a-bc5e-40d0f80974e5" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.042015] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a8805827-63d1-4694-b9e2-a3130943470a tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.188s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.066272] env[63531]: INFO nova.compute.manager [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Rebuilding instance [ 883.104344] env[63531]: DEBUG oslo_vmware.api [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1118064, 'name': CreateSnapshot_Task, 'duration_secs': 1.210995} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.104763] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Created Snapshot of the VM instance {{(pid=63531) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 883.105596] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04a0863d-4972-4449-a580-b0a1ab19210c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.116344] env[63531]: DEBUG nova.compute.manager [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 883.116344] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe0ded36-305b-4e70-a513-35b11338c51d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.127035] env[63531]: DEBUG nova.compute.utils [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 883.133018] env[63531]: DEBUG nova.compute.manager [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 883.133018] env[63531]: DEBUG nova.network.neutron [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 883.215331] env[63531]: DEBUG nova.policy [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '36cb2f0de54f417384fa1fb2210de4e4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8ba8b284854242c392aec5326e996239', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 883.462670] env[63531]: DEBUG nova.compute.manager [req-a48fc242-3afa-4a8a-9c63-edba1f51a297 req-3e61a8a4-09a7-4ba3-a07b-4dc483414826 service nova] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Received event network-changed-073adca7-1360-4703-9eb1-d86758bb8002 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 883.462884] env[63531]: DEBUG nova.compute.manager [req-a48fc242-3afa-4a8a-9c63-edba1f51a297 req-3e61a8a4-09a7-4ba3-a07b-4dc483414826 service nova] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Refreshing instance network info cache due to event network-changed-073adca7-1360-4703-9eb1-d86758bb8002. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 883.463122] env[63531]: DEBUG oslo_concurrency.lockutils [req-a48fc242-3afa-4a8a-9c63-edba1f51a297 req-3e61a8a4-09a7-4ba3-a07b-4dc483414826 service nova] Acquiring lock "refresh_cache-37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.463273] env[63531]: DEBUG oslo_concurrency.lockutils [req-a48fc242-3afa-4a8a-9c63-edba1f51a297 req-3e61a8a4-09a7-4ba3-a07b-4dc483414826 service nova] Acquired lock "refresh_cache-37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.463435] env[63531]: DEBUG nova.network.neutron [req-a48fc242-3afa-4a8a-9c63-edba1f51a297 req-3e61a8a4-09a7-4ba3-a07b-4dc483414826 service nova] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Refreshing network info cache for port 073adca7-1360-4703-9eb1-d86758bb8002 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 883.477994] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e54157c-b2af-4c09-ad30-d81f45506a37 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 883.478280] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-00094479-48ba-42a7-98a4-aba1ada3cc3b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.487975] env[63531]: DEBUG oslo_vmware.api [None req-7e54157c-b2af-4c09-ad30-d81f45506a37 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 883.487975] env[63531]: value = "task-1118065" [ 883.487975] env[63531]: _type = "Task" [ 883.487975] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.492362] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-0365d10b-233e-47b3-a10a-1ec120c16583 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 883.492642] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7d6a70d7-9d53-494f-b6c8-927644b4b4f6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.501390] env[63531]: DEBUG oslo_vmware.api [None req-7e54157c-b2af-4c09-ad30-d81f45506a37 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118065, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.503064] env[63531]: DEBUG oslo_vmware.api [None req-0365d10b-233e-47b3-a10a-1ec120c16583 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Waiting for the task: (returnval){ [ 883.503064] env[63531]: value = "task-1118066" [ 883.503064] env[63531]: _type = "Task" [ 883.503064] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.512698] env[63531]: DEBUG oslo_vmware.api [None req-0365d10b-233e-47b3-a10a-1ec120c16583 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118066, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.597947] env[63531]: DEBUG nova.network.neutron [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Successfully created port: 17da826c-920c-4044-93cb-be8c8cf01c13 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 883.635072] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Creating linked-clone VM from snapshot {{(pid=63531) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 883.635894] env[63531]: DEBUG nova.compute.manager [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 883.641287] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-0ff6e09c-cf0a-4b21-a2bc-9788b4a3266f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.646019] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 883.646337] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9a572830-59ad-41ce-921b-280ccf64e241 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.656440] env[63531]: DEBUG oslo_vmware.api [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Waiting for the task: (returnval){ [ 883.656440] env[63531]: value = "task-1118068" [ 883.656440] env[63531]: _type = "Task" [ 883.656440] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.661219] env[63531]: DEBUG oslo_vmware.api [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 883.661219] env[63531]: value = "task-1118067" [ 883.661219] env[63531]: _type = "Task" [ 883.661219] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.675716] env[63531]: DEBUG oslo_vmware.api [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1118068, 'name': CloneVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.683237] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] VM already powered off {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 883.683589] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 883.684580] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bc9abf7-5d62-4918-b528-ecb1952fc620 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.701385] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 883.701746] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5a3c27b9-d10d-4be8-a592-63fd93a39960 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.799666] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 883.800075] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 883.800204] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Deleting the datastore file [datastore1] e95d3273-8216-47cc-95b6-99301366a827 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 883.801251] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6856453e-9278-4999-b8b8-717a7bd438b2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.809971] env[63531]: DEBUG oslo_vmware.api [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 883.809971] env[63531]: value = "task-1118070" [ 883.809971] env[63531]: _type = "Task" [ 883.809971] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.827602] env[63531]: DEBUG oslo_vmware.api [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118070, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.005713] env[63531]: DEBUG oslo_vmware.api [None req-7e54157c-b2af-4c09-ad30-d81f45506a37 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118065, 'name': PowerOffVM_Task, 'duration_secs': 0.280978} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.009371] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e54157c-b2af-4c09-ad30-d81f45506a37 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 884.012484] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3236e46c-5e64-4f50-afd1-b12214bc1ad2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.024250] env[63531]: DEBUG oslo_vmware.api [None req-0365d10b-233e-47b3-a10a-1ec120c16583 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118066, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.044382] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0053fecc-9481-489d-b80f-65326ee18cb2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.086095] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a201177-31e0-4b97-9e6f-8b945d2e797a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.090198] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e54157c-b2af-4c09-ad30-d81f45506a37 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 884.090439] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-de49ef53-72f7-4ff2-9395-e0944fd8c732 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.095238] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05e86b63-08bb-4d3d-84f1-85031bb55ade {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.103148] env[63531]: DEBUG oslo_vmware.api [None req-7e54157c-b2af-4c09-ad30-d81f45506a37 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 884.103148] env[63531]: value = "task-1118071" [ 884.103148] env[63531]: _type = "Task" [ 884.103148] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.137532] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af5d2ea2-d737-46ea-9408-56b2e2c8963f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.144850] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e54157c-b2af-4c09-ad30-d81f45506a37 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] VM already powered off {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 884.145092] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-7e54157c-b2af-4c09-ad30-d81f45506a37 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 884.145416] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7e54157c-b2af-4c09-ad30-d81f45506a37 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.145597] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7e54157c-b2af-4c09-ad30-d81f45506a37 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.145788] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e54157c-b2af-4c09-ad30-d81f45506a37 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 884.150020] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0e28fe67-4a6b-4c09-9d88-4674f0482adb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.158502] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dacdd185-39b3-4152-acff-7aa8e1306586 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.167864] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e54157c-b2af-4c09-ad30-d81f45506a37 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 884.168139] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-7e54157c-b2af-4c09-ad30-d81f45506a37 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 884.169481] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a7184845-51e6-4957-856b-65d783e002a5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.184396] env[63531]: DEBUG nova.compute.provider_tree [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 884.185909] env[63531]: DEBUG oslo_vmware.api [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1118068, 'name': CloneVM_Task} progress is 94%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.193371] env[63531]: DEBUG oslo_vmware.api [None req-7e54157c-b2af-4c09-ad30-d81f45506a37 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 884.193371] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52265aef-55bc-b39b-155d-e696675c09f2" [ 884.193371] env[63531]: _type = "Task" [ 884.193371] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.205335] env[63531]: DEBUG oslo_vmware.api [None req-7e54157c-b2af-4c09-ad30-d81f45506a37 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52265aef-55bc-b39b-155d-e696675c09f2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.326057] env[63531]: DEBUG oslo_vmware.api [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118070, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.174566} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.326215] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 884.326521] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 884.327495] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 884.450644] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5cd60179-ca4c-47ad-8ed1-0bb74872da0a tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Acquiring lock "45b05c54-6865-47a1-ba93-90ad3e1ba07e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.450803] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5cd60179-ca4c-47ad-8ed1-0bb74872da0a tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Lock "45b05c54-6865-47a1-ba93-90ad3e1ba07e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.451504] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5cd60179-ca4c-47ad-8ed1-0bb74872da0a tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Acquiring lock "45b05c54-6865-47a1-ba93-90ad3e1ba07e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.451504] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5cd60179-ca4c-47ad-8ed1-0bb74872da0a tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Lock "45b05c54-6865-47a1-ba93-90ad3e1ba07e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.451504] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5cd60179-ca4c-47ad-8ed1-0bb74872da0a tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Lock "45b05c54-6865-47a1-ba93-90ad3e1ba07e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.455291] env[63531]: INFO nova.compute.manager [None req-5cd60179-ca4c-47ad-8ed1-0bb74872da0a tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] Terminating instance [ 884.457517] env[63531]: DEBUG nova.compute.manager [None req-5cd60179-ca4c-47ad-8ed1-0bb74872da0a tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 884.457758] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5cd60179-ca4c-47ad-8ed1-0bb74872da0a tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 884.458763] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8361907-8957-4a8e-801c-9be229666a03 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.469975] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-5cd60179-ca4c-47ad-8ed1-0bb74872da0a tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 884.470323] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-74c74376-fca2-4afa-a2c8-d29032f5a2b6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.480164] env[63531]: DEBUG oslo_vmware.api [None req-5cd60179-ca4c-47ad-8ed1-0bb74872da0a tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Waiting for the task: (returnval){ [ 884.480164] env[63531]: value = "task-1118072" [ 884.480164] env[63531]: _type = "Task" [ 884.480164] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.491423] env[63531]: DEBUG oslo_vmware.api [None req-5cd60179-ca4c-47ad-8ed1-0bb74872da0a tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Task: {'id': task-1118072, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.516747] env[63531]: DEBUG oslo_vmware.api [None req-0365d10b-233e-47b3-a10a-1ec120c16583 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118066, 'name': PowerOnVM_Task, 'duration_secs': 0.625223} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.517109] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-0365d10b-233e-47b3-a10a-1ec120c16583 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 884.517333] env[63531]: DEBUG nova.compute.manager [None req-0365d10b-233e-47b3-a10a-1ec120c16583 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 884.518262] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d50a1f6b-4eae-44e5-afd5-e24767f3b34c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.618136] env[63531]: DEBUG nova.network.neutron [req-a48fc242-3afa-4a8a-9c63-edba1f51a297 req-3e61a8a4-09a7-4ba3-a07b-4dc483414826 service nova] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Updated VIF entry in instance network info cache for port 073adca7-1360-4703-9eb1-d86758bb8002. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 884.618571] env[63531]: DEBUG nova.network.neutron [req-a48fc242-3afa-4a8a-9c63-edba1f51a297 req-3e61a8a4-09a7-4ba3-a07b-4dc483414826 service nova] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Updating instance_info_cache with network_info: [{"id": "073adca7-1360-4703-9eb1-d86758bb8002", "address": "fa:16:3e:9b:a3:5c", "network": {"id": "95ed2c7d-3d1a-4736-8e90-104002f5c40b", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-73549344-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9056486e3b640d0bcdb9369ed5d67ad", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04e15990-16e1-4cb2-b0f0-06c362e68c5e", "external-id": "nsx-vlan-transportzone-555", "segmentation_id": 555, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap073adca7-13", "ovs_interfaceid": "073adca7-1360-4703-9eb1-d86758bb8002", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.656580] env[63531]: DEBUG nova.compute.manager [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 884.674334] env[63531]: DEBUG oslo_vmware.api [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1118068, 'name': CloneVM_Task} progress is 94%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.685358] env[63531]: DEBUG nova.virt.hardware [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 884.685612] env[63531]: DEBUG nova.virt.hardware [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 884.685789] env[63531]: DEBUG nova.virt.hardware [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 884.685994] env[63531]: DEBUG nova.virt.hardware [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 884.685994] env[63531]: DEBUG nova.virt.hardware [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 884.686355] env[63531]: DEBUG nova.virt.hardware [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 884.686535] env[63531]: DEBUG nova.virt.hardware [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 884.686699] env[63531]: DEBUG nova.virt.hardware [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 884.686909] env[63531]: DEBUG nova.virt.hardware [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 884.687101] env[63531]: DEBUG nova.virt.hardware [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 884.687416] env[63531]: DEBUG nova.virt.hardware [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 884.688320] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37aad766-b92a-4386-8070-63f9a1f7785a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.691729] env[63531]: DEBUG nova.scheduler.client.report [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 884.708788] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24f50ea0-260b-426b-8074-7f8027a58124 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.728030] env[63531]: DEBUG oslo_vmware.api [None req-7e54157c-b2af-4c09-ad30-d81f45506a37 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52265aef-55bc-b39b-155d-e696675c09f2, 'name': SearchDatastore_Task, 'duration_secs': 0.012303} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.728030] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ca4f64ef-ac12-4ed0-8804-4a4c2e9cc0b0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.744370] env[63531]: DEBUG oslo_vmware.api [None req-7e54157c-b2af-4c09-ad30-d81f45506a37 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 884.744370] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]525d7e2e-071c-710b-9b87-50849948b913" [ 884.744370] env[63531]: _type = "Task" [ 884.744370] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.756240] env[63531]: DEBUG oslo_vmware.api [None req-7e54157c-b2af-4c09-ad30-d81f45506a37 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]525d7e2e-071c-710b-9b87-50849948b913, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.996386] env[63531]: DEBUG oslo_vmware.api [None req-5cd60179-ca4c-47ad-8ed1-0bb74872da0a tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Task: {'id': task-1118072, 'name': PowerOffVM_Task, 'duration_secs': 0.40486} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.996386] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-5cd60179-ca4c-47ad-8ed1-0bb74872da0a tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 884.996386] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5cd60179-ca4c-47ad-8ed1-0bb74872da0a tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 884.996386] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-62822d5c-2909-4bb4-a20b-5b4320a6c20d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.094793] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5cd60179-ca4c-47ad-8ed1-0bb74872da0a tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 885.094793] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5cd60179-ca4c-47ad-8ed1-0bb74872da0a tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 885.094793] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-5cd60179-ca4c-47ad-8ed1-0bb74872da0a tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Deleting the datastore file [datastore1] 45b05c54-6865-47a1-ba93-90ad3e1ba07e {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 885.095118] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2c834e44-1bab-4bd9-9d24-1e35dd3a9df7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.104757] env[63531]: DEBUG oslo_vmware.api [None req-5cd60179-ca4c-47ad-8ed1-0bb74872da0a tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Waiting for the task: (returnval){ [ 885.104757] env[63531]: value = "task-1118074" [ 885.104757] env[63531]: _type = "Task" [ 885.104757] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.118269] env[63531]: DEBUG oslo_vmware.api [None req-5cd60179-ca4c-47ad-8ed1-0bb74872da0a tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Task: {'id': task-1118074, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.122135] env[63531]: DEBUG oslo_concurrency.lockutils [req-a48fc242-3afa-4a8a-9c63-edba1f51a297 req-3e61a8a4-09a7-4ba3-a07b-4dc483414826 service nova] Releasing lock "refresh_cache-37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.173573] env[63531]: DEBUG oslo_vmware.api [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1118068, 'name': CloneVM_Task} progress is 95%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.209509] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.590s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.210018] env[63531]: DEBUG nova.compute.manager [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 885.218492] env[63531]: DEBUG oslo_concurrency.lockutils [None req-36b75f90-354a-4910-9550-00922114a1d3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.897s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.218824] env[63531]: DEBUG nova.objects.instance [None req-36b75f90-354a-4910-9550-00922114a1d3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lazy-loading 'resources' on Instance uuid 2b0e4285-310a-439c-bc50-000a7c5ef7f9 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 885.261981] env[63531]: DEBUG oslo_vmware.api [None req-7e54157c-b2af-4c09-ad30-d81f45506a37 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]525d7e2e-071c-710b-9b87-50849948b913, 'name': SearchDatastore_Task, 'duration_secs': 0.011774} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.262319] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7e54157c-b2af-4c09-ad30-d81f45506a37 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.262590] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e54157c-b2af-4c09-ad30-d81f45506a37 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df/9d550399-588c-4bd5-8436-eba9b523bd13-rescue.vmdk. {{(pid=63531) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 885.262867] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-75b5d6c9-50dd-47de-9c63-bf016e287d43 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.270813] env[63531]: DEBUG oslo_vmware.api [None req-7e54157c-b2af-4c09-ad30-d81f45506a37 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 885.270813] env[63531]: value = "task-1118075" [ 885.270813] env[63531]: _type = "Task" [ 885.270813] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.279854] env[63531]: DEBUG oslo_vmware.api [None req-7e54157c-b2af-4c09-ad30-d81f45506a37 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118075, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.376269] env[63531]: DEBUG nova.virt.hardware [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 885.376269] env[63531]: DEBUG nova.virt.hardware [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 885.376442] env[63531]: DEBUG nova.virt.hardware [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 885.376563] env[63531]: DEBUG nova.virt.hardware [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 885.376772] env[63531]: DEBUG nova.virt.hardware [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 885.377571] env[63531]: DEBUG nova.virt.hardware [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 885.377571] env[63531]: DEBUG nova.virt.hardware [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 885.377716] env[63531]: DEBUG nova.virt.hardware [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 885.377907] env[63531]: DEBUG nova.virt.hardware [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 885.378104] env[63531]: DEBUG nova.virt.hardware [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 885.378312] env[63531]: DEBUG nova.virt.hardware [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 885.379263] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24f724f3-37e3-428e-ae49-d48fe46194b8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.389908] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db13d03b-3789-4be6-bf93-f8607c368e26 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.407305] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1e:bc:64', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'db068f71-08cc-42d4-8ab6-17134c1585e5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a806c2a6-5a84-4114-aedb-8bb2da53a147', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 885.415692] env[63531]: DEBUG oslo.service.loopingcall [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 885.416051] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e95d3273-8216-47cc-95b6-99301366a827] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 885.416289] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f57ce6cc-e22c-4b30-beb3-4de81878154e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.438915] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 885.438915] env[63531]: value = "task-1118076" [ 885.438915] env[63531]: _type = "Task" [ 885.438915] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.452609] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118076, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.616366] env[63531]: DEBUG oslo_vmware.api [None req-5cd60179-ca4c-47ad-8ed1-0bb74872da0a tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Task: {'id': task-1118074, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.147527} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.616986] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-5cd60179-ca4c-47ad-8ed1-0bb74872da0a tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 885.617446] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5cd60179-ca4c-47ad-8ed1-0bb74872da0a tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 885.617760] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5cd60179-ca4c-47ad-8ed1-0bb74872da0a tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 885.618087] env[63531]: INFO nova.compute.manager [None req-5cd60179-ca4c-47ad-8ed1-0bb74872da0a tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] Took 1.16 seconds to destroy the instance on the hypervisor. [ 885.618476] env[63531]: DEBUG oslo.service.loopingcall [None req-5cd60179-ca4c-47ad-8ed1-0bb74872da0a tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 885.618815] env[63531]: DEBUG nova.compute.manager [-] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 885.619028] env[63531]: DEBUG nova.network.neutron [-] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 885.676845] env[63531]: DEBUG oslo_vmware.api [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1118068, 'name': CloneVM_Task, 'duration_secs': 1.710869} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.677972] env[63531]: INFO nova.virt.vmwareapi.vmops [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Created linked-clone VM from snapshot [ 885.679181] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f02d0f1-5b20-4419-8669-5ff1260f9703 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.694349] env[63531]: DEBUG nova.virt.vmwareapi.images [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Uploading image d2fca167-8767-4b16-b64a-61fd11e58a0e {{(pid=63531) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 885.716927] env[63531]: DEBUG oslo_vmware.rw_handles [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 885.716927] env[63531]: value = "vm-244730" [ 885.716927] env[63531]: _type = "VirtualMachine" [ 885.716927] env[63531]: }. {{(pid=63531) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 885.717112] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-1d962684-1d93-4797-84e0-a0abf88829a0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.724318] env[63531]: DEBUG nova.compute.utils [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 885.730035] env[63531]: DEBUG nova.compute.manager [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 885.730035] env[63531]: DEBUG nova.network.neutron [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 885.739808] env[63531]: DEBUG oslo_vmware.rw_handles [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Lease: (returnval){ [ 885.739808] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52edea11-8adf-9f62-40b8-7cb26cf645ab" [ 885.739808] env[63531]: _type = "HttpNfcLease" [ 885.739808] env[63531]: } obtained for exporting VM: (result){ [ 885.739808] env[63531]: value = "vm-244730" [ 885.739808] env[63531]: _type = "VirtualMachine" [ 885.739808] env[63531]: }. {{(pid=63531) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 885.740179] env[63531]: DEBUG oslo_vmware.api [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Waiting for the lease: (returnval){ [ 885.740179] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52edea11-8adf-9f62-40b8-7cb26cf645ab" [ 885.740179] env[63531]: _type = "HttpNfcLease" [ 885.740179] env[63531]: } to be ready. {{(pid=63531) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 885.754611] env[63531]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 885.754611] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52edea11-8adf-9f62-40b8-7cb26cf645ab" [ 885.754611] env[63531]: _type = "HttpNfcLease" [ 885.754611] env[63531]: } is initializing. {{(pid=63531) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 885.788356] env[63531]: DEBUG oslo_vmware.api [None req-7e54157c-b2af-4c09-ad30-d81f45506a37 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118075, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.504361} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.789036] env[63531]: INFO nova.virt.vmwareapi.ds_util [None req-7e54157c-b2af-4c09-ad30-d81f45506a37 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df/9d550399-588c-4bd5-8436-eba9b523bd13-rescue.vmdk. [ 885.789611] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dda150e5-197d-4a3a-8643-a8505bebbbe7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.793472] env[63531]: DEBUG nova.policy [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '049cb89a769f41bd91d0e62784635cbd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c2e99349303e413b9f7896d449cb7dff', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 885.822157] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-7e54157c-b2af-4c09-ad30-d81f45506a37 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Reconfiguring VM instance instance-00000046 to attach disk [datastore1] 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df/9d550399-588c-4bd5-8436-eba9b523bd13-rescue.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 885.825890] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-93f7ef4c-e1cc-428e-b5e2-30150a2266e8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.846351] env[63531]: DEBUG oslo_vmware.api [None req-7e54157c-b2af-4c09-ad30-d81f45506a37 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 885.846351] env[63531]: value = "task-1118078" [ 885.846351] env[63531]: _type = "Task" [ 885.846351] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.856084] env[63531]: DEBUG oslo_vmware.api [None req-7e54157c-b2af-4c09-ad30-d81f45506a37 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118078, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.933178] env[63531]: DEBUG nova.network.neutron [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Successfully updated port: 17da826c-920c-4044-93cb-be8c8cf01c13 {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 885.955301] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118076, 'name': CreateVM_Task} progress is 25%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.978012] env[63531]: DEBUG oslo_vmware.rw_handles [None req-89c9bc13-a8c9-4b1a-9244-7179ff01bbf6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/522b4d34-0b98-d713-f9e1-190e2d0581fd/disk-0.vmdk. {{(pid=63531) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 885.979121] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34747851-6259-4fa2-ad76-d6660943fb9e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.985747] env[63531]: DEBUG oslo_vmware.rw_handles [None req-89c9bc13-a8c9-4b1a-9244-7179ff01bbf6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/522b4d34-0b98-d713-f9e1-190e2d0581fd/disk-0.vmdk is in state: ready. {{(pid=63531) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 885.985889] env[63531]: ERROR oslo_vmware.rw_handles [None req-89c9bc13-a8c9-4b1a-9244-7179ff01bbf6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/522b4d34-0b98-d713-f9e1-190e2d0581fd/disk-0.vmdk due to incomplete transfer. [ 885.988823] env[63531]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-8311d1bb-1a38-48e4-b640-37cd81b4fca2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.999190] env[63531]: DEBUG oslo_vmware.rw_handles [None req-89c9bc13-a8c9-4b1a-9244-7179ff01bbf6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/522b4d34-0b98-d713-f9e1-190e2d0581fd/disk-0.vmdk. {{(pid=63531) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 885.999502] env[63531]: DEBUG nova.virt.vmwareapi.images [None req-89c9bc13-a8c9-4b1a-9244-7179ff01bbf6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Uploaded image f998b6ba-c103-40c9-a8fa-cecaa1f6d66c to the Glance image server {{(pid=63531) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 886.001680] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-89c9bc13-a8c9-4b1a-9244-7179ff01bbf6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Destroying the VM {{(pid=63531) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 886.001984] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-9ff8c3c0-8f3f-4026-85ef-e1c6a23087c1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.013358] env[63531]: DEBUG oslo_vmware.api [None req-89c9bc13-a8c9-4b1a-9244-7179ff01bbf6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 886.013358] env[63531]: value = "task-1118079" [ 886.013358] env[63531]: _type = "Task" [ 886.013358] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.025076] env[63531]: DEBUG oslo_vmware.api [None req-89c9bc13-a8c9-4b1a-9244-7179ff01bbf6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118079, 'name': Destroy_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.145999] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a87c658-ad31-4c6e-81d5-317729bba775 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.154480] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f27f5e43-076c-4a58-b90e-788405a785cf {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.193741] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ed99b40-5801-4123-b30c-99b3e794a934 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.201132] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4496fe01-b5d0-4cc6-a23d-3b8c896f9a4d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.221479] env[63531]: DEBUG nova.compute.provider_tree [None req-36b75f90-354a-4910-9550-00922114a1d3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 886.230654] env[63531]: DEBUG nova.compute.manager [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 886.249448] env[63531]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 886.249448] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52edea11-8adf-9f62-40b8-7cb26cf645ab" [ 886.249448] env[63531]: _type = "HttpNfcLease" [ 886.249448] env[63531]: } is ready. {{(pid=63531) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 886.249866] env[63531]: DEBUG oslo_vmware.rw_handles [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 886.249866] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52edea11-8adf-9f62-40b8-7cb26cf645ab" [ 886.249866] env[63531]: _type = "HttpNfcLease" [ 886.249866] env[63531]: }. {{(pid=63531) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 886.251134] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad608c4e-835d-4efe-89df-f17d13ad550f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.260474] env[63531]: DEBUG oslo_vmware.rw_handles [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f97289-5f27-b8d5-8c36-8b0cbd3d7139/disk-0.vmdk from lease info. {{(pid=63531) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 886.261679] env[63531]: DEBUG oslo_vmware.rw_handles [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f97289-5f27-b8d5-8c36-8b0cbd3d7139/disk-0.vmdk for reading. {{(pid=63531) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 886.325378] env[63531]: DEBUG nova.network.neutron [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Successfully created port: 3e1c2a04-40a1-4e53-8770-dd615bd761eb {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 886.328112] env[63531]: DEBUG nova.compute.manager [req-7f7eb923-ee0f-4a48-a7b6-02c32f16eab1 req-35a2a428-296f-44d4-aff2-c3b067fdb21e service nova] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Received event network-vif-plugged-17da826c-920c-4044-93cb-be8c8cf01c13 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 886.328357] env[63531]: DEBUG oslo_concurrency.lockutils [req-7f7eb923-ee0f-4a48-a7b6-02c32f16eab1 req-35a2a428-296f-44d4-aff2-c3b067fdb21e service nova] Acquiring lock "a090135f-489d-457d-be36-ba6f61b71ab8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.328612] env[63531]: DEBUG oslo_concurrency.lockutils [req-7f7eb923-ee0f-4a48-a7b6-02c32f16eab1 req-35a2a428-296f-44d4-aff2-c3b067fdb21e service nova] Lock "a090135f-489d-457d-be36-ba6f61b71ab8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.328808] env[63531]: DEBUG oslo_concurrency.lockutils [req-7f7eb923-ee0f-4a48-a7b6-02c32f16eab1 req-35a2a428-296f-44d4-aff2-c3b067fdb21e service nova] Lock "a090135f-489d-457d-be36-ba6f61b71ab8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.329130] env[63531]: DEBUG nova.compute.manager [req-7f7eb923-ee0f-4a48-a7b6-02c32f16eab1 req-35a2a428-296f-44d4-aff2-c3b067fdb21e service nova] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] No waiting events found dispatching network-vif-plugged-17da826c-920c-4044-93cb-be8c8cf01c13 {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 886.329334] env[63531]: WARNING nova.compute.manager [req-7f7eb923-ee0f-4a48-a7b6-02c32f16eab1 req-35a2a428-296f-44d4-aff2-c3b067fdb21e service nova] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Received unexpected event network-vif-plugged-17da826c-920c-4044-93cb-be8c8cf01c13 for instance with vm_state building and task_state spawning. [ 886.357738] env[63531]: DEBUG oslo_vmware.api [None req-7e54157c-b2af-4c09-ad30-d81f45506a37 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118078, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.364149] env[63531]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-e8854edc-cde0-4098-83f3-d63fb09a88a4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.439893] env[63531]: DEBUG oslo_concurrency.lockutils [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "refresh_cache-a090135f-489d-457d-be36-ba6f61b71ab8" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.440078] env[63531]: DEBUG oslo_concurrency.lockutils [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquired lock "refresh_cache-a090135f-489d-457d-be36-ba6f61b71ab8" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.440256] env[63531]: DEBUG nova.network.neutron [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 886.454993] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118076, 'name': CreateVM_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.530892] env[63531]: DEBUG oslo_vmware.api [None req-89c9bc13-a8c9-4b1a-9244-7179ff01bbf6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118079, 'name': Destroy_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.723351] env[63531]: DEBUG nova.scheduler.client.report [None req-36b75f90-354a-4910-9550-00922114a1d3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 886.845068] env[63531]: DEBUG nova.network.neutron [-] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 886.876227] env[63531]: DEBUG oslo_vmware.api [None req-7e54157c-b2af-4c09-ad30-d81f45506a37 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118078, 'name': ReconfigVM_Task, 'duration_secs': 0.581064} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.877891] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-7e54157c-b2af-4c09-ad30-d81f45506a37 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Reconfigured VM instance instance-00000046 to attach disk [datastore1] 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df/9d550399-588c-4bd5-8436-eba9b523bd13-rescue.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 886.881265] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5daad30e-0d37-46cb-abf3-a748b370866e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.926240] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-17d3c0a2-3623-445b-b5d1-a5a973d94320 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.939806] env[63531]: DEBUG nova.compute.manager [req-b4d250e9-642a-434c-8633-e583a48d0e4a req-29d60c52-73e7-4596-9a71-0a80099f7144 service nova] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] Received event network-vif-deleted-e93a4c61-805f-4c14-8649-6718ca8f490d {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 886.955787] env[63531]: DEBUG oslo_vmware.api [None req-7e54157c-b2af-4c09-ad30-d81f45506a37 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 886.955787] env[63531]: value = "task-1118080" [ 886.955787] env[63531]: _type = "Task" [ 886.955787] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.960053] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118076, 'name': CreateVM_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.970800] env[63531]: DEBUG oslo_vmware.api [None req-7e54157c-b2af-4c09-ad30-d81f45506a37 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118080, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.027321] env[63531]: DEBUG oslo_vmware.api [None req-89c9bc13-a8c9-4b1a-9244-7179ff01bbf6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118079, 'name': Destroy_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.035609] env[63531]: DEBUG nova.network.neutron [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 887.229835] env[63531]: DEBUG oslo_concurrency.lockutils [None req-36b75f90-354a-4910-9550-00922114a1d3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.011s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.232650] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6a5546ec-a279-4010-b37e-f0e5bb9ca4fe tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.320s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.233117] env[63531]: DEBUG nova.objects.instance [None req-6a5546ec-a279-4010-b37e-f0e5bb9ca4fe tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Lazy-loading 'resources' on Instance uuid db83adf4-6183-4330-b260-77d1f5daf899 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 887.246622] env[63531]: DEBUG nova.compute.manager [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 887.261800] env[63531]: INFO nova.scheduler.client.report [None req-36b75f90-354a-4910-9550-00922114a1d3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Deleted allocations for instance 2b0e4285-310a-439c-bc50-000a7c5ef7f9 [ 887.282178] env[63531]: DEBUG nova.virt.hardware [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 887.282382] env[63531]: DEBUG nova.virt.hardware [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 887.282510] env[63531]: DEBUG nova.virt.hardware [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 887.282771] env[63531]: DEBUG nova.virt.hardware [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 887.284138] env[63531]: DEBUG nova.virt.hardware [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 887.284138] env[63531]: DEBUG nova.virt.hardware [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 887.284138] env[63531]: DEBUG nova.virt.hardware [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 887.284138] env[63531]: DEBUG nova.virt.hardware [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 887.284358] env[63531]: DEBUG nova.virt.hardware [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 887.284973] env[63531]: DEBUG nova.virt.hardware [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 887.284973] env[63531]: DEBUG nova.virt.hardware [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 887.286863] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e146fd1-5687-4e21-a0f5-11c55785b0e8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.298468] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-027b98be-7b76-4347-a488-3e307440cf94 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.325017] env[63531]: DEBUG nova.network.neutron [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Updating instance_info_cache with network_info: [{"id": "17da826c-920c-4044-93cb-be8c8cf01c13", "address": "fa:16:3e:9a:ac:ad", "network": {"id": "caf86428-59a5-4e50-9135-2ab8473af64d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1673994696-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ba8b284854242c392aec5326e996239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap17da826c-92", "ovs_interfaceid": "17da826c-920c-4044-93cb-be8c8cf01c13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.347533] env[63531]: INFO nova.compute.manager [-] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] Took 1.73 seconds to deallocate network for instance. [ 887.455888] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118076, 'name': CreateVM_Task, 'duration_secs': 1.686664} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.456212] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e95d3273-8216-47cc-95b6-99301366a827] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 887.457038] env[63531]: DEBUG oslo_concurrency.lockutils [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.457256] env[63531]: DEBUG oslo_concurrency.lockutils [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.457643] env[63531]: DEBUG oslo_concurrency.lockutils [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 887.458107] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c2b96a7d-0bfc-4da6-90ea-aa40d6665db3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.472256] env[63531]: DEBUG oslo_vmware.api [None req-7e54157c-b2af-4c09-ad30-d81f45506a37 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118080, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.473441] env[63531]: DEBUG oslo_vmware.api [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 887.473441] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52d7f532-fbd3-5baa-f32a-abddf0cbf164" [ 887.473441] env[63531]: _type = "Task" [ 887.473441] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.483971] env[63531]: DEBUG oslo_vmware.api [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52d7f532-fbd3-5baa-f32a-abddf0cbf164, 'name': SearchDatastore_Task, 'duration_secs': 0.01129} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.484331] env[63531]: DEBUG oslo_concurrency.lockutils [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.484570] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 887.485206] env[63531]: DEBUG oslo_concurrency.lockutils [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.485395] env[63531]: DEBUG oslo_concurrency.lockutils [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.485576] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 887.485866] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-03c28355-d029-431d-b563-645f4ef3fbba {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.499239] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "6118e242-c0db-4707-a5f9-3d12b823935f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.499593] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "6118e242-c0db-4707-a5f9-3d12b823935f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.503063] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 887.503063] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 887.503063] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f1ad24a9-c8d9-45e3-8b1c-f3a0a69f29e3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.510370] env[63531]: DEBUG oslo_vmware.api [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 887.510370] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5236448e-b27e-346d-e580-77ef41982a82" [ 887.510370] env[63531]: _type = "Task" [ 887.510370] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.532769] env[63531]: DEBUG oslo_vmware.api [None req-89c9bc13-a8c9-4b1a-9244-7179ff01bbf6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118079, 'name': Destroy_Task} progress is 33%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.533088] env[63531]: DEBUG oslo_vmware.api [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5236448e-b27e-346d-e580-77ef41982a82, 'name': SearchDatastore_Task, 'duration_secs': 0.011084} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.537263] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-66755e9b-d0d3-4764-9801-fcdfabccedc0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.542516] env[63531]: DEBUG oslo_vmware.api [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 887.542516] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52bd3ff4-a932-fc25-6bf6-1288cb06557c" [ 887.542516] env[63531]: _type = "Task" [ 887.542516] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.553372] env[63531]: DEBUG oslo_vmware.api [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52bd3ff4-a932-fc25-6bf6-1288cb06557c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.776657] env[63531]: DEBUG oslo_concurrency.lockutils [None req-36b75f90-354a-4910-9550-00922114a1d3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "2b0e4285-310a-439c-bc50-000a7c5ef7f9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.558s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.829993] env[63531]: DEBUG oslo_concurrency.lockutils [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Releasing lock "refresh_cache-a090135f-489d-457d-be36-ba6f61b71ab8" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.829993] env[63531]: DEBUG nova.compute.manager [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Instance network_info: |[{"id": "17da826c-920c-4044-93cb-be8c8cf01c13", "address": "fa:16:3e:9a:ac:ad", "network": {"id": "caf86428-59a5-4e50-9135-2ab8473af64d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1673994696-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ba8b284854242c392aec5326e996239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap17da826c-92", "ovs_interfaceid": "17da826c-920c-4044-93cb-be8c8cf01c13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 887.829993] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9a:ac:ad', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '91b0f7e5-0d1a-46e2-bf73-09656211dea2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '17da826c-920c-4044-93cb-be8c8cf01c13', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 887.839131] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Creating folder: Project (8ba8b284854242c392aec5326e996239). Parent ref: group-v244585. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 887.842533] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d0de22f1-92a8-4508-aa78-893448231f7d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.855759] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5cd60179-ca4c-47ad-8ed1-0bb74872da0a tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.860570] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Created folder: Project (8ba8b284854242c392aec5326e996239) in parent group-v244585. [ 887.861094] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Creating folder: Instances. Parent ref: group-v244732. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 887.862604] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f8c29392-0c58-4567-bdf6-e07f9e64bf85 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.875212] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Created folder: Instances in parent group-v244732. [ 887.875621] env[63531]: DEBUG oslo.service.loopingcall [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 887.875970] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 887.876288] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7ae12355-bcfc-467f-80e3-30df91340666 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.904766] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 887.904766] env[63531]: value = "task-1118083" [ 887.904766] env[63531]: _type = "Task" [ 887.904766] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.915911] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118083, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.986716] env[63531]: DEBUG oslo_vmware.api [None req-7e54157c-b2af-4c09-ad30-d81f45506a37 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118080, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.008051] env[63531]: DEBUG nova.compute.manager [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 888.029315] env[63531]: DEBUG oslo_vmware.api [None req-89c9bc13-a8c9-4b1a-9244-7179ff01bbf6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118079, 'name': Destroy_Task, 'duration_secs': 1.749786} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.029315] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-89c9bc13-a8c9-4b1a-9244-7179ff01bbf6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Destroyed the VM [ 888.029315] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-89c9bc13-a8c9-4b1a-9244-7179ff01bbf6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Deleting Snapshot of the VM instance {{(pid=63531) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 888.029315] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-48394b29-6f7d-4326-bff5-2f77c617da85 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.041700] env[63531]: DEBUG oslo_vmware.api [None req-89c9bc13-a8c9-4b1a-9244-7179ff01bbf6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 888.041700] env[63531]: value = "task-1118084" [ 888.041700] env[63531]: _type = "Task" [ 888.041700] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.060574] env[63531]: DEBUG oslo_vmware.api [None req-89c9bc13-a8c9-4b1a-9244-7179ff01bbf6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118084, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.066475] env[63531]: DEBUG oslo_vmware.api [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52bd3ff4-a932-fc25-6bf6-1288cb06557c, 'name': SearchDatastore_Task, 'duration_secs': 0.010272} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.066876] env[63531]: DEBUG oslo_concurrency.lockutils [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.067312] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] e95d3273-8216-47cc-95b6-99301366a827/e95d3273-8216-47cc-95b6-99301366a827.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 888.067701] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-36f36dea-b1f5-4724-8fd6-25e6704a2a49 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.081135] env[63531]: DEBUG oslo_vmware.api [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 888.081135] env[63531]: value = "task-1118085" [ 888.081135] env[63531]: _type = "Task" [ 888.081135] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.098208] env[63531]: DEBUG oslo_vmware.api [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118085, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.165968] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc632912-43d8-4173-892a-8813ee56aba1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.176677] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-131b79b4-4599-4832-80e2-ddbb207d571c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.214139] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39434552-530b-4ab1-88ff-4d2c85ffa617 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.223299] env[63531]: DEBUG nova.network.neutron [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Successfully updated port: 3e1c2a04-40a1-4e53-8770-dd615bd761eb {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 888.225781] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfd68e6e-a4e5-46ce-a444-94b5b2b4a30f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.249108] env[63531]: DEBUG nova.compute.provider_tree [None req-6a5546ec-a279-4010-b37e-f0e5bb9ca4fe tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 888.316540] env[63531]: DEBUG nova.compute.manager [req-6a72a37d-c1b7-450a-b5b8-dd43cd4a3cfa req-cff44239-37d2-4676-ad06-e95b2c6ce8cf service nova] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Received event network-changed-17da826c-920c-4044-93cb-be8c8cf01c13 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 888.316827] env[63531]: DEBUG nova.compute.manager [req-6a72a37d-c1b7-450a-b5b8-dd43cd4a3cfa req-cff44239-37d2-4676-ad06-e95b2c6ce8cf service nova] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Refreshing instance network info cache due to event network-changed-17da826c-920c-4044-93cb-be8c8cf01c13. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 888.317229] env[63531]: DEBUG oslo_concurrency.lockutils [req-6a72a37d-c1b7-450a-b5b8-dd43cd4a3cfa req-cff44239-37d2-4676-ad06-e95b2c6ce8cf service nova] Acquiring lock "refresh_cache-a090135f-489d-457d-be36-ba6f61b71ab8" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.317693] env[63531]: DEBUG oslo_concurrency.lockutils [req-6a72a37d-c1b7-450a-b5b8-dd43cd4a3cfa req-cff44239-37d2-4676-ad06-e95b2c6ce8cf service nova] Acquired lock "refresh_cache-a090135f-489d-457d-be36-ba6f61b71ab8" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.317943] env[63531]: DEBUG nova.network.neutron [req-6a72a37d-c1b7-450a-b5b8-dd43cd4a3cfa req-cff44239-37d2-4676-ad06-e95b2c6ce8cf service nova] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Refreshing network info cache for port 17da826c-920c-4044-93cb-be8c8cf01c13 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 888.420528] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118083, 'name': CreateVM_Task, 'duration_secs': 0.490732} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.423104] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 888.423104] env[63531]: DEBUG oslo_concurrency.lockutils [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.423104] env[63531]: DEBUG oslo_concurrency.lockutils [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.423104] env[63531]: DEBUG oslo_concurrency.lockutils [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 888.423104] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b23fc98c-c17b-4ea0-b4eb-1f3992cf3ff8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.431125] env[63531]: DEBUG oslo_vmware.api [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for the task: (returnval){ [ 888.431125] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52be0450-8346-e5fb-603b-b0f8efb1d51b" [ 888.431125] env[63531]: _type = "Task" [ 888.431125] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.445229] env[63531]: DEBUG oslo_vmware.api [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52be0450-8346-e5fb-603b-b0f8efb1d51b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.476194] env[63531]: DEBUG oslo_vmware.api [None req-7e54157c-b2af-4c09-ad30-d81f45506a37 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118080, 'name': ReconfigVM_Task, 'duration_secs': 1.302574} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.476586] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e54157c-b2af-4c09-ad30-d81f45506a37 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 888.476980] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bbca0762-4ea7-48df-9c9f-40a9e2e9d62d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.485611] env[63531]: DEBUG oslo_vmware.api [None req-7e54157c-b2af-4c09-ad30-d81f45506a37 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 888.485611] env[63531]: value = "task-1118086" [ 888.485611] env[63531]: _type = "Task" [ 888.485611] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.496470] env[63531]: DEBUG oslo_vmware.api [None req-7e54157c-b2af-4c09-ad30-d81f45506a37 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118086, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.533871] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.555709] env[63531]: DEBUG oslo_vmware.api [None req-89c9bc13-a8c9-4b1a-9244-7179ff01bbf6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118084, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.597192] env[63531]: DEBUG oslo_vmware.api [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118085, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.732545] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "refresh_cache-dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.732741] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquired lock "refresh_cache-dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.732901] env[63531]: DEBUG nova.network.neutron [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 888.755047] env[63531]: DEBUG nova.scheduler.client.report [None req-6a5546ec-a279-4010-b37e-f0e5bb9ca4fe tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 888.943756] env[63531]: DEBUG oslo_vmware.api [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52be0450-8346-e5fb-603b-b0f8efb1d51b, 'name': SearchDatastore_Task, 'duration_secs': 0.046898} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.943756] env[63531]: DEBUG oslo_concurrency.lockutils [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.943943] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 888.944189] env[63531]: DEBUG oslo_concurrency.lockutils [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.944342] env[63531]: DEBUG oslo_concurrency.lockutils [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.944539] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 888.944814] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-36a07e6e-b804-40be-b75b-d36c14d8aeac {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.955505] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 888.955913] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 888.956681] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aca0a0d4-9f2f-455e-9cb8-973fcf0dcddc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.963927] env[63531]: DEBUG oslo_vmware.api [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for the task: (returnval){ [ 888.963927] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52acebab-5974-1b5f-c419-74f7ec86a3d5" [ 888.963927] env[63531]: _type = "Task" [ 888.963927] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.975305] env[63531]: DEBUG oslo_vmware.api [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52acebab-5974-1b5f-c419-74f7ec86a3d5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.996648] env[63531]: DEBUG oslo_vmware.api [None req-7e54157c-b2af-4c09-ad30-d81f45506a37 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118086, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.058462] env[63531]: DEBUG oslo_vmware.api [None req-89c9bc13-a8c9-4b1a-9244-7179ff01bbf6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118084, 'name': RemoveSnapshot_Task, 'duration_secs': 0.787102} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.058772] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-89c9bc13-a8c9-4b1a-9244-7179ff01bbf6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Deleted Snapshot of the VM instance {{(pid=63531) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 889.062034] env[63531]: INFO nova.compute.manager [None req-89c9bc13-a8c9-4b1a-9244-7179ff01bbf6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Took 17.35 seconds to snapshot the instance on the hypervisor. [ 889.099310] env[63531]: DEBUG oslo_vmware.api [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118085, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.556678} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.099556] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] e95d3273-8216-47cc-95b6-99301366a827/e95d3273-8216-47cc-95b6-99301366a827.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 889.099808] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 889.100818] env[63531]: DEBUG nova.network.neutron [req-6a72a37d-c1b7-450a-b5b8-dd43cd4a3cfa req-cff44239-37d2-4676-ad06-e95b2c6ce8cf service nova] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Updated VIF entry in instance network info cache for port 17da826c-920c-4044-93cb-be8c8cf01c13. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 889.101162] env[63531]: DEBUG nova.network.neutron [req-6a72a37d-c1b7-450a-b5b8-dd43cd4a3cfa req-cff44239-37d2-4676-ad06-e95b2c6ce8cf service nova] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Updating instance_info_cache with network_info: [{"id": "17da826c-920c-4044-93cb-be8c8cf01c13", "address": "fa:16:3e:9a:ac:ad", "network": {"id": "caf86428-59a5-4e50-9135-2ab8473af64d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1673994696-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ba8b284854242c392aec5326e996239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap17da826c-92", "ovs_interfaceid": "17da826c-920c-4044-93cb-be8c8cf01c13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.102362] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d9a4c9eb-4473-489e-843c-e50984cfbd87 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.115346] env[63531]: DEBUG oslo_vmware.api [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 889.115346] env[63531]: value = "task-1118087" [ 889.115346] env[63531]: _type = "Task" [ 889.115346] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.131990] env[63531]: DEBUG oslo_vmware.api [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118087, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.237073] env[63531]: DEBUG nova.compute.manager [req-d5c79205-7383-4143-b54f-714dd15eeea2 req-98479e32-6a35-40cc-b8f0-b5265a0ed77f service nova] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Received event network-vif-plugged-3e1c2a04-40a1-4e53-8770-dd615bd761eb {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 889.237319] env[63531]: DEBUG oslo_concurrency.lockutils [req-d5c79205-7383-4143-b54f-714dd15eeea2 req-98479e32-6a35-40cc-b8f0-b5265a0ed77f service nova] Acquiring lock "dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.237729] env[63531]: DEBUG oslo_concurrency.lockutils [req-d5c79205-7383-4143-b54f-714dd15eeea2 req-98479e32-6a35-40cc-b8f0-b5265a0ed77f service nova] Lock "dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.239451] env[63531]: DEBUG oslo_concurrency.lockutils [req-d5c79205-7383-4143-b54f-714dd15eeea2 req-98479e32-6a35-40cc-b8f0-b5265a0ed77f service nova] Lock "dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.239451] env[63531]: DEBUG nova.compute.manager [req-d5c79205-7383-4143-b54f-714dd15eeea2 req-98479e32-6a35-40cc-b8f0-b5265a0ed77f service nova] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] No waiting events found dispatching network-vif-plugged-3e1c2a04-40a1-4e53-8770-dd615bd761eb {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 889.239451] env[63531]: WARNING nova.compute.manager [req-d5c79205-7383-4143-b54f-714dd15eeea2 req-98479e32-6a35-40cc-b8f0-b5265a0ed77f service nova] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Received unexpected event network-vif-plugged-3e1c2a04-40a1-4e53-8770-dd615bd761eb for instance with vm_state building and task_state spawning. [ 889.239451] env[63531]: DEBUG nova.compute.manager [req-d5c79205-7383-4143-b54f-714dd15eeea2 req-98479e32-6a35-40cc-b8f0-b5265a0ed77f service nova] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Received event network-changed-3e1c2a04-40a1-4e53-8770-dd615bd761eb {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 889.239451] env[63531]: DEBUG nova.compute.manager [req-d5c79205-7383-4143-b54f-714dd15eeea2 req-98479e32-6a35-40cc-b8f0-b5265a0ed77f service nova] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Refreshing instance network info cache due to event network-changed-3e1c2a04-40a1-4e53-8770-dd615bd761eb. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 889.239451] env[63531]: DEBUG oslo_concurrency.lockutils [req-d5c79205-7383-4143-b54f-714dd15eeea2 req-98479e32-6a35-40cc-b8f0-b5265a0ed77f service nova] Acquiring lock "refresh_cache-dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 889.260946] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6a5546ec-a279-4010-b37e-f0e5bb9ca4fe tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.028s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.268491] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.712s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.268491] env[63531]: INFO nova.compute.claims [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 889.284911] env[63531]: DEBUG nova.network.neutron [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 889.313323] env[63531]: INFO nova.scheduler.client.report [None req-6a5546ec-a279-4010-b37e-f0e5bb9ca4fe tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Deleted allocations for instance db83adf4-6183-4330-b260-77d1f5daf899 [ 889.480436] env[63531]: DEBUG oslo_vmware.api [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52acebab-5974-1b5f-c419-74f7ec86a3d5, 'name': SearchDatastore_Task, 'duration_secs': 0.01226} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.481756] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f333daf1-835d-4887-bd91-68c2bf3926e4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.486246] env[63531]: DEBUG nova.network.neutron [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Updating instance_info_cache with network_info: [{"id": "3e1c2a04-40a1-4e53-8770-dd615bd761eb", "address": "fa:16:3e:a5:e0:6e", "network": {"id": "bf99b8cb-9efb-4a74-9db1-0f5f0875f26a", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-998205595-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2e99349303e413b9f7896d449cb7dff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc6e6fe1-c4f5-4389-a49f-0978060eebb4", "external-id": "nsx-vlan-transportzone-829", "segmentation_id": 829, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e1c2a04-40", "ovs_interfaceid": "3e1c2a04-40a1-4e53-8770-dd615bd761eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.493256] env[63531]: DEBUG oslo_vmware.api [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for the task: (returnval){ [ 889.493256] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5243f545-2c21-d101-97da-b5fea1de897f" [ 889.493256] env[63531]: _type = "Task" [ 889.493256] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.502969] env[63531]: DEBUG oslo_vmware.api [None req-7e54157c-b2af-4c09-ad30-d81f45506a37 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118086, 'name': PowerOnVM_Task, 'duration_secs': 0.537889} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.503948] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e54157c-b2af-4c09-ad30-d81f45506a37 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 889.509865] env[63531]: DEBUG oslo_vmware.api [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5243f545-2c21-d101-97da-b5fea1de897f, 'name': SearchDatastore_Task, 'duration_secs': 0.011823} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.511018] env[63531]: DEBUG nova.compute.manager [None req-7e54157c-b2af-4c09-ad30-d81f45506a37 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 889.511490] env[63531]: DEBUG oslo_concurrency.lockutils [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.511650] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] a090135f-489d-457d-be36-ba6f61b71ab8/a090135f-489d-457d-be36-ba6f61b71ab8.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 889.512566] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c328509b-7ce2-4d72-bb25-6fe0bf51d76a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.515934] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f241bddf-2c6e-4ca2-9246-b9b4af95d760 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.527920] env[63531]: DEBUG oslo_vmware.api [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for the task: (returnval){ [ 889.527920] env[63531]: value = "task-1118088" [ 889.527920] env[63531]: _type = "Task" [ 889.527920] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.539230] env[63531]: DEBUG oslo_vmware.api [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118088, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.610219] env[63531]: DEBUG nova.compute.manager [None req-89c9bc13-a8c9-4b1a-9244-7179ff01bbf6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Found 3 images (rotation: 2) {{(pid=63531) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 889.610458] env[63531]: DEBUG nova.compute.manager [None req-89c9bc13-a8c9-4b1a-9244-7179ff01bbf6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Rotating out 1 backups {{(pid=63531) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4562}} [ 889.610660] env[63531]: DEBUG nova.compute.manager [None req-89c9bc13-a8c9-4b1a-9244-7179ff01bbf6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Deleting image 4f765107-1431-4888-adf7-e8b9dbcda251 {{(pid=63531) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4567}} [ 889.612940] env[63531]: DEBUG oslo_concurrency.lockutils [req-6a72a37d-c1b7-450a-b5b8-dd43cd4a3cfa req-cff44239-37d2-4676-ad06-e95b2c6ce8cf service nova] Releasing lock "refresh_cache-a090135f-489d-457d-be36-ba6f61b71ab8" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.629500] env[63531]: DEBUG oslo_vmware.api [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118087, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.147165} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.629838] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 889.630734] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49e9f669-77ab-4c8b-9f9c-f6070e3b83bf {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.657632] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Reconfiguring VM instance instance-00000048 to attach disk [datastore1] e95d3273-8216-47cc-95b6-99301366a827/e95d3273-8216-47cc-95b6-99301366a827.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 889.658060] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fe975b99-8853-47cc-9411-5308e5139ad2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.680611] env[63531]: DEBUG oslo_vmware.api [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 889.680611] env[63531]: value = "task-1118089" [ 889.680611] env[63531]: _type = "Task" [ 889.680611] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.695183] env[63531]: DEBUG oslo_vmware.api [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118089, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.823239] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6a5546ec-a279-4010-b37e-f0e5bb9ca4fe tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Lock "db83adf4-6183-4330-b260-77d1f5daf899" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.074s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.995799] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Releasing lock "refresh_cache-dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.996165] env[63531]: DEBUG nova.compute.manager [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Instance network_info: |[{"id": "3e1c2a04-40a1-4e53-8770-dd615bd761eb", "address": "fa:16:3e:a5:e0:6e", "network": {"id": "bf99b8cb-9efb-4a74-9db1-0f5f0875f26a", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-998205595-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2e99349303e413b9f7896d449cb7dff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc6e6fe1-c4f5-4389-a49f-0978060eebb4", "external-id": "nsx-vlan-transportzone-829", "segmentation_id": 829, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e1c2a04-40", "ovs_interfaceid": "3e1c2a04-40a1-4e53-8770-dd615bd761eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 889.996683] env[63531]: DEBUG oslo_concurrency.lockutils [req-d5c79205-7383-4143-b54f-714dd15eeea2 req-98479e32-6a35-40cc-b8f0-b5265a0ed77f service nova] Acquired lock "refresh_cache-dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 889.996779] env[63531]: DEBUG nova.network.neutron [req-d5c79205-7383-4143-b54f-714dd15eeea2 req-98479e32-6a35-40cc-b8f0-b5265a0ed77f service nova] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Refreshing network info cache for port 3e1c2a04-40a1-4e53-8770-dd615bd761eb {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 890.001747] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a5:e0:6e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bc6e6fe1-c4f5-4389-a49f-0978060eebb4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3e1c2a04-40a1-4e53-8770-dd615bd761eb', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 890.008673] env[63531]: DEBUG oslo.service.loopingcall [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 890.010340] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 890.010594] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9f69b186-a2e9-4171-b6e9-ff11646b89ce {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.047526] env[63531]: DEBUG oslo_vmware.api [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118088, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.507745} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.049052] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] a090135f-489d-457d-be36-ba6f61b71ab8/a090135f-489d-457d-be36-ba6f61b71ab8.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 890.049304] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 890.049885] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 890.049885] env[63531]: value = "task-1118090" [ 890.049885] env[63531]: _type = "Task" [ 890.049885] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.050265] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d8e72233-2087-40c1-8f85-01c86b7dd0d1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.064754] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118090, 'name': CreateVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.066674] env[63531]: DEBUG oslo_vmware.api [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for the task: (returnval){ [ 890.066674] env[63531]: value = "task-1118091" [ 890.066674] env[63531]: _type = "Task" [ 890.066674] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.075751] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "06ec91b2-14cb-4db8-8fab-e6c4183b1116" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.076043] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "06ec91b2-14cb-4db8-8fab-e6c4183b1116" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.082208] env[63531]: DEBUG oslo_vmware.api [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118091, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.193050] env[63531]: DEBUG oslo_vmware.api [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118089, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.444285] env[63531]: DEBUG oslo_concurrency.lockutils [None req-817aa9b4-fcda-44b3-b34f-da4ac7528014 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Acquiring lock "1e5e28ef-38ce-47be-ac36-8bdf17ab0a47" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.444610] env[63531]: DEBUG oslo_concurrency.lockutils [None req-817aa9b4-fcda-44b3-b34f-da4ac7528014 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Lock "1e5e28ef-38ce-47be-ac36-8bdf17ab0a47" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.444798] env[63531]: DEBUG oslo_concurrency.lockutils [None req-817aa9b4-fcda-44b3-b34f-da4ac7528014 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Acquiring lock "1e5e28ef-38ce-47be-ac36-8bdf17ab0a47-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.444893] env[63531]: DEBUG oslo_concurrency.lockutils [None req-817aa9b4-fcda-44b3-b34f-da4ac7528014 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Lock "1e5e28ef-38ce-47be-ac36-8bdf17ab0a47-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.445078] env[63531]: DEBUG oslo_concurrency.lockutils [None req-817aa9b4-fcda-44b3-b34f-da4ac7528014 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Lock "1e5e28ef-38ce-47be-ac36-8bdf17ab0a47-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.449589] env[63531]: INFO nova.compute.manager [None req-817aa9b4-fcda-44b3-b34f-da4ac7528014 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Terminating instance [ 890.450499] env[63531]: DEBUG nova.compute.manager [None req-817aa9b4-fcda-44b3-b34f-da4ac7528014 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 890.450835] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-817aa9b4-fcda-44b3-b34f-da4ac7528014 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 890.451695] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40924b4e-777e-40f9-b5a8-620ea4ef0c51 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.462369] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-817aa9b4-fcda-44b3-b34f-da4ac7528014 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 890.462666] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4d66cb09-1e2b-4c01-b141-8aebc5afc9b1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.472684] env[63531]: DEBUG oslo_vmware.api [None req-817aa9b4-fcda-44b3-b34f-da4ac7528014 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Waiting for the task: (returnval){ [ 890.472684] env[63531]: value = "task-1118092" [ 890.472684] env[63531]: _type = "Task" [ 890.472684] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.484208] env[63531]: DEBUG oslo_vmware.api [None req-817aa9b4-fcda-44b3-b34f-da4ac7528014 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1118092, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.572302] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118090, 'name': CreateVM_Task} progress is 25%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.583505] env[63531]: DEBUG nova.compute.manager [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 06ec91b2-14cb-4db8-8fab-e6c4183b1116] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 890.596516] env[63531]: DEBUG oslo_vmware.api [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118091, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078368} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.600598] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 890.603026] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f3266b1-0e94-4faa-b8d3-4382097de16b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.632644] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Reconfiguring VM instance instance-0000004a to attach disk [datastore1] a090135f-489d-457d-be36-ba6f61b71ab8/a090135f-489d-457d-be36-ba6f61b71ab8.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 890.638029] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1094c348-b2df-4a00-943b-33bd64ddc9a4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.665387] env[63531]: DEBUG oslo_vmware.api [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for the task: (returnval){ [ 890.665387] env[63531]: value = "task-1118093" [ 890.665387] env[63531]: _type = "Task" [ 890.665387] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.679273] env[63531]: DEBUG oslo_vmware.api [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118093, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.682325] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d150cc09-f880-4b9c-8508-32be7d1b1367 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.699297] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a16f3726-cdfc-47a9-98e0-9f8b901b2361 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.703231] env[63531]: DEBUG oslo_vmware.api [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118089, 'name': ReconfigVM_Task, 'duration_secs': 0.671913} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.703231] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Reconfigured VM instance instance-00000048 to attach disk [datastore1] e95d3273-8216-47cc-95b6-99301366a827/e95d3273-8216-47cc-95b6-99301366a827.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 890.704207] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3143e4d2-4dad-4000-9be8-f494ec81a1a6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.738837] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1664a40f-1908-4981-8c77-15390ae0550e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.741653] env[63531]: DEBUG oslo_vmware.api [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 890.741653] env[63531]: value = "task-1118094" [ 890.741653] env[63531]: _type = "Task" [ 890.741653] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.749910] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5aae29c4-7de7-48e7-b6df-a7684b5d7d13 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.759264] env[63531]: DEBUG oslo_vmware.api [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118094, 'name': Rename_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.771621] env[63531]: DEBUG nova.compute.provider_tree [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 890.856307] env[63531]: DEBUG nova.network.neutron [req-d5c79205-7383-4143-b54f-714dd15eeea2 req-98479e32-6a35-40cc-b8f0-b5265a0ed77f service nova] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Updated VIF entry in instance network info cache for port 3e1c2a04-40a1-4e53-8770-dd615bd761eb. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 890.856708] env[63531]: DEBUG nova.network.neutron [req-d5c79205-7383-4143-b54f-714dd15eeea2 req-98479e32-6a35-40cc-b8f0-b5265a0ed77f service nova] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Updating instance_info_cache with network_info: [{"id": "3e1c2a04-40a1-4e53-8770-dd615bd761eb", "address": "fa:16:3e:a5:e0:6e", "network": {"id": "bf99b8cb-9efb-4a74-9db1-0f5f0875f26a", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-998205595-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2e99349303e413b9f7896d449cb7dff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc6e6fe1-c4f5-4389-a49f-0978060eebb4", "external-id": "nsx-vlan-transportzone-829", "segmentation_id": 829, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e1c2a04-40", "ovs_interfaceid": "3e1c2a04-40a1-4e53-8770-dd615bd761eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 890.985041] env[63531]: DEBUG oslo_vmware.api [None req-817aa9b4-fcda-44b3-b34f-da4ac7528014 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1118092, 'name': PowerOffVM_Task, 'duration_secs': 0.311374} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.985041] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-817aa9b4-fcda-44b3-b34f-da4ac7528014 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 890.985270] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-817aa9b4-fcda-44b3-b34f-da4ac7528014 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 890.985977] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-866fa077-e2ef-421e-8c49-e252eb535a42 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.063322] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-817aa9b4-fcda-44b3-b34f-da4ac7528014 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 891.063322] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-817aa9b4-fcda-44b3-b34f-da4ac7528014 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 891.063322] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-817aa9b4-fcda-44b3-b34f-da4ac7528014 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Deleting the datastore file [datastore1] 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 891.067598] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-99d9b6ee-a9fc-47f3-9b2a-4eb12a9214ef {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.070202] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118090, 'name': CreateVM_Task, 'duration_secs': 0.724611} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.073323] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 891.073323] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.073323] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.073323] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 891.073323] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f04ec0e9-b51b-4819-bc08-8096b78a748f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.077234] env[63531]: DEBUG oslo_vmware.api [None req-817aa9b4-fcda-44b3-b34f-da4ac7528014 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Waiting for the task: (returnval){ [ 891.077234] env[63531]: value = "task-1118096" [ 891.077234] env[63531]: _type = "Task" [ 891.077234] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.083375] env[63531]: DEBUG oslo_vmware.api [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 891.083375] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52583802-5cf6-0e56-3ee8-b581da27d46c" [ 891.083375] env[63531]: _type = "Task" [ 891.083375] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.097800] env[63531]: DEBUG oslo_vmware.api [None req-817aa9b4-fcda-44b3-b34f-da4ac7528014 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1118096, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.107810] env[63531]: DEBUG oslo_vmware.api [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52583802-5cf6-0e56-3ee8-b581da27d46c, 'name': SearchDatastore_Task, 'duration_secs': 0.017568} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.108133] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.108652] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 891.108652] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.108849] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.109234] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 891.109302] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ee496d7f-605e-4e24-815e-ed585c6d16fb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.118956] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.120629] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 891.120847] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 891.121869] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-895a25de-8ea6-49ef-b705-9dc1ff4639b6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.129508] env[63531]: DEBUG oslo_vmware.api [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 891.129508] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52d0f32e-2fb3-8f1e-85f9-20ea06da8e67" [ 891.129508] env[63531]: _type = "Task" [ 891.129508] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.141356] env[63531]: DEBUG oslo_vmware.api [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52d0f32e-2fb3-8f1e-85f9-20ea06da8e67, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.178614] env[63531]: DEBUG oslo_vmware.api [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118093, 'name': ReconfigVM_Task, 'duration_secs': 0.373593} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.179088] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Reconfigured VM instance instance-0000004a to attach disk [datastore1] a090135f-489d-457d-be36-ba6f61b71ab8/a090135f-489d-457d-be36-ba6f61b71ab8.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 891.180960] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b20b30c7-f8dc-471a-a926-36d0ef4f4891 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.185996] env[63531]: DEBUG oslo_concurrency.lockutils [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Acquiring lock "4b5e5821-9def-4b53-be19-355d9e5f81ec" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.186581] env[63531]: DEBUG oslo_concurrency.lockutils [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Lock "4b5e5821-9def-4b53-be19-355d9e5f81ec" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.195548] env[63531]: DEBUG oslo_vmware.api [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for the task: (returnval){ [ 891.195548] env[63531]: value = "task-1118097" [ 891.195548] env[63531]: _type = "Task" [ 891.195548] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.207147] env[63531]: DEBUG oslo_vmware.api [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118097, 'name': Rename_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.257795] env[63531]: DEBUG oslo_vmware.api [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118094, 'name': Rename_Task, 'duration_secs': 0.244315} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.257795] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 891.258084] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-48190f91-6737-45a0-afa4-299456112a91 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.266912] env[63531]: DEBUG oslo_vmware.api [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 891.266912] env[63531]: value = "task-1118098" [ 891.266912] env[63531]: _type = "Task" [ 891.266912] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.278455] env[63531]: DEBUG nova.scheduler.client.report [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 891.282737] env[63531]: DEBUG oslo_vmware.api [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118098, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.360581] env[63531]: DEBUG oslo_concurrency.lockutils [req-d5c79205-7383-4143-b54f-714dd15eeea2 req-98479e32-6a35-40cc-b8f0-b5265a0ed77f service nova] Releasing lock "refresh_cache-dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.589859] env[63531]: DEBUG oslo_vmware.api [None req-817aa9b4-fcda-44b3-b34f-da4ac7528014 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1118096, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.290015} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.589859] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-817aa9b4-fcda-44b3-b34f-da4ac7528014 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 891.590318] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-817aa9b4-fcda-44b3-b34f-da4ac7528014 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 891.590318] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-817aa9b4-fcda-44b3-b34f-da4ac7528014 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 891.590469] env[63531]: INFO nova.compute.manager [None req-817aa9b4-fcda-44b3-b34f-da4ac7528014 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Took 1.14 seconds to destroy the instance on the hypervisor. [ 891.590718] env[63531]: DEBUG oslo.service.loopingcall [None req-817aa9b4-fcda-44b3-b34f-da4ac7528014 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 891.590925] env[63531]: DEBUG nova.compute.manager [-] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 891.591032] env[63531]: DEBUG nova.network.neutron [-] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 891.599635] env[63531]: INFO nova.compute.manager [None req-9e75f160-ec7f-4a3b-9503-b6838e2ef42a tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Unrescuing [ 891.600035] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9e75f160-ec7f-4a3b-9503-b6838e2ef42a tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquiring lock "refresh_cache-4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.600131] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9e75f160-ec7f-4a3b-9503-b6838e2ef42a tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquired lock "refresh_cache-4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.600341] env[63531]: DEBUG nova.network.neutron [None req-9e75f160-ec7f-4a3b-9503-b6838e2ef42a tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 891.644528] env[63531]: DEBUG oslo_vmware.api [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52d0f32e-2fb3-8f1e-85f9-20ea06da8e67, 'name': SearchDatastore_Task, 'duration_secs': 0.019219} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.645424] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e2a736e9-0df8-4109-a028-7e0d7fba8c16 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.651763] env[63531]: DEBUG oslo_vmware.api [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 891.651763] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5242a033-ad50-9d62-2caf-8d7cef63aba2" [ 891.651763] env[63531]: _type = "Task" [ 891.651763] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.660781] env[63531]: DEBUG oslo_vmware.api [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5242a033-ad50-9d62-2caf-8d7cef63aba2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.691027] env[63531]: DEBUG nova.compute.manager [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 891.706834] env[63531]: DEBUG oslo_vmware.api [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118097, 'name': Rename_Task, 'duration_secs': 0.242935} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.708405] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 891.708405] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ceb1f11e-8298-49e4-ad0b-c181d1af64ac {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.716022] env[63531]: DEBUG oslo_vmware.api [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for the task: (returnval){ [ 891.716022] env[63531]: value = "task-1118099" [ 891.716022] env[63531]: _type = "Task" [ 891.716022] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.725686] env[63531]: DEBUG oslo_vmware.api [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118099, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.777920] env[63531]: DEBUG oslo_vmware.api [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118098, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.786046] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.521s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.786656] env[63531]: DEBUG nova.compute.manager [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 891.793019] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8a9dc91a-31e6-4fb7-b47e-0b940db4e5dc tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.076s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.793019] env[63531]: DEBUG nova.objects.instance [None req-8a9dc91a-31e6-4fb7-b47e-0b940db4e5dc tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Lazy-loading 'resources' on Instance uuid e7fb00a3-0d2e-4f54-950d-337307112d7a {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 892.078903] env[63531]: DEBUG nova.compute.manager [req-4ea35620-2054-46d6-bd54-374da0d164fd req-5c16e150-cdc2-4b0b-8387-c8f4a531c124 service nova] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Received event network-vif-deleted-d4b7599b-ca73-4bab-91d4-33a977545afc {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 892.078974] env[63531]: INFO nova.compute.manager [req-4ea35620-2054-46d6-bd54-374da0d164fd req-5c16e150-cdc2-4b0b-8387-c8f4a531c124 service nova] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Neutron deleted interface d4b7599b-ca73-4bab-91d4-33a977545afc; detaching it from the instance and deleting it from the info cache [ 892.079259] env[63531]: DEBUG nova.network.neutron [req-4ea35620-2054-46d6-bd54-374da0d164fd req-5c16e150-cdc2-4b0b-8387-c8f4a531c124 service nova] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.164640] env[63531]: DEBUG oslo_vmware.api [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5242a033-ad50-9d62-2caf-8d7cef63aba2, 'name': SearchDatastore_Task, 'duration_secs': 0.030808} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.164942] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.165308] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a/dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 892.165610] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-667b1047-d5c0-44b6-a9c7-29e40dabe1af {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.174721] env[63531]: DEBUG oslo_vmware.api [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 892.174721] env[63531]: value = "task-1118100" [ 892.174721] env[63531]: _type = "Task" [ 892.174721] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.185628] env[63531]: DEBUG oslo_vmware.api [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118100, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.214666] env[63531]: DEBUG oslo_concurrency.lockutils [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.227638] env[63531]: DEBUG oslo_vmware.api [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118099, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.276833] env[63531]: DEBUG oslo_vmware.api [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118098, 'name': PowerOnVM_Task, 'duration_secs': 0.800553} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.277120] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 892.277344] env[63531]: DEBUG nova.compute.manager [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 892.278179] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdb3abc4-80ac-4e00-bda3-74779e5af921 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.296676] env[63531]: DEBUG nova.compute.utils [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 892.305945] env[63531]: DEBUG nova.compute.manager [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 892.305945] env[63531]: DEBUG nova.network.neutron [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 892.369505] env[63531]: DEBUG nova.network.neutron [None req-9e75f160-ec7f-4a3b-9503-b6838e2ef42a tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Updating instance_info_cache with network_info: [{"id": "e30711c3-e609-434b-82da-a58252d3678d", "address": "fa:16:3e:54:c1:b7", "network": {"id": "913045d9-e90b-4366-987c-7706b73a6147", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1696510389-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f71c7db0a3614995becc25d9a8dce787", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c297fe21-cd0b-4226-813b-a65d2358d034", "external-id": "nsx-vlan-transportzone-98", "segmentation_id": 98, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape30711c3-e6", "ovs_interfaceid": "e30711c3-e609-434b-82da-a58252d3678d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.377869] env[63531]: DEBUG nova.policy [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7140e12106504b3e95ff49786361771b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '63900312a30b4942854f552758c039cf', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 892.399650] env[63531]: DEBUG nova.network.neutron [-] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.581969] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ec110462-f38a-4ff3-9ba2-a44e39693e01 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.594429] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dc1fbb6-3c7d-4e44-99be-632411d66491 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.636370] env[63531]: DEBUG nova.compute.manager [req-4ea35620-2054-46d6-bd54-374da0d164fd req-5c16e150-cdc2-4b0b-8387-c8f4a531c124 service nova] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Detach interface failed, port_id=d4b7599b-ca73-4bab-91d4-33a977545afc, reason: Instance 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47 could not be found. {{(pid=63531) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 892.688424] env[63531]: DEBUG oslo_vmware.api [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118100, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.704976] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e1a9d51-3374-4de7-845e-cf5208ca4bc4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.715826] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d95fe9c1-e47a-4ca2-b81d-6596016cecd1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.736184] env[63531]: DEBUG oslo_vmware.api [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118099, 'name': PowerOnVM_Task, 'duration_secs': 0.625071} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.759528] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 892.759760] env[63531]: INFO nova.compute.manager [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Took 8.10 seconds to spawn the instance on the hypervisor. [ 892.760115] env[63531]: DEBUG nova.compute.manager [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 892.761031] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b220d31-7058-447b-9131-8f917f7937f5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.764248] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15200297-1ed1-409e-9244-c2cec85cd992 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.768116] env[63531]: DEBUG nova.network.neutron [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] Successfully created port: 161e87a1-1133-476f-9598-44c29595c36a {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 892.777596] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a0660b8-4133-4c28-a712-f7ee07b1d75d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.797462] env[63531]: INFO nova.compute.manager [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] bringing vm to original state: 'stopped' [ 892.804023] env[63531]: DEBUG nova.compute.provider_tree [None req-8a9dc91a-31e6-4fb7-b47e-0b940db4e5dc tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 892.805176] env[63531]: DEBUG nova.compute.manager [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 892.817082] env[63531]: DEBUG oslo_concurrency.lockutils [None req-2b52d1c6-91c9-4f04-ad9c-74a98d502f8d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquiring lock "9a3fad46-a15a-451c-bdab-a3c8cc8add07" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.817677] env[63531]: DEBUG oslo_concurrency.lockutils [None req-2b52d1c6-91c9-4f04-ad9c-74a98d502f8d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "9a3fad46-a15a-451c-bdab-a3c8cc8add07" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.818378] env[63531]: DEBUG nova.compute.manager [None req-2b52d1c6-91c9-4f04-ad9c-74a98d502f8d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 892.819094] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a837109c-3f29-4ccd-ac15-0cb0953c406d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.829279] env[63531]: DEBUG nova.compute.manager [None req-2b52d1c6-91c9-4f04-ad9c-74a98d502f8d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63531) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 892.829796] env[63531]: DEBUG nova.objects.instance [None req-2b52d1c6-91c9-4f04-ad9c-74a98d502f8d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lazy-loading 'flavor' on Instance uuid 9a3fad46-a15a-451c-bdab-a3c8cc8add07 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 892.875188] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9e75f160-ec7f-4a3b-9503-b6838e2ef42a tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Releasing lock "refresh_cache-4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.875932] env[63531]: DEBUG nova.objects.instance [None req-9e75f160-ec7f-4a3b-9503-b6838e2ef42a tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Lazy-loading 'flavor' on Instance uuid 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 892.903031] env[63531]: INFO nova.compute.manager [-] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Took 1.31 seconds to deallocate network for instance. [ 893.188128] env[63531]: DEBUG oslo_vmware.api [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118100, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.645335} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.188433] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a/dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 893.188672] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 893.188957] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0596837e-7654-43f3-8ac1-e00b285dde62 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.196232] env[63531]: DEBUG oslo_vmware.api [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 893.196232] env[63531]: value = "task-1118101" [ 893.196232] env[63531]: _type = "Task" [ 893.196232] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.205599] env[63531]: DEBUG oslo_vmware.api [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118101, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.297128] env[63531]: INFO nova.compute.manager [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Took 22.05 seconds to build instance. [ 893.305632] env[63531]: DEBUG nova.scheduler.client.report [None req-8a9dc91a-31e6-4fb7-b47e-0b940db4e5dc tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 893.336904] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b52d1c6-91c9-4f04-ad9c-74a98d502f8d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 893.337221] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-87e067fd-133b-4e24-987c-6a933584bcd9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.345442] env[63531]: DEBUG oslo_vmware.api [None req-2b52d1c6-91c9-4f04-ad9c-74a98d502f8d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 893.345442] env[63531]: value = "task-1118102" [ 893.345442] env[63531]: _type = "Task" [ 893.345442] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.356300] env[63531]: DEBUG oslo_vmware.api [None req-2b52d1c6-91c9-4f04-ad9c-74a98d502f8d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118102, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.383577] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fa5451e-2b9e-4886-b429-c2f30997b8e1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.406359] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e75f160-ec7f-4a3b-9503-b6838e2ef42a tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 893.406724] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ae742fb2-49db-4350-8bb8-fec925eba3d8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.409273] env[63531]: DEBUG oslo_concurrency.lockutils [None req-817aa9b4-fcda-44b3-b34f-da4ac7528014 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.416144] env[63531]: DEBUG oslo_vmware.api [None req-9e75f160-ec7f-4a3b-9503-b6838e2ef42a tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 893.416144] env[63531]: value = "task-1118103" [ 893.416144] env[63531]: _type = "Task" [ 893.416144] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.425049] env[63531]: DEBUG oslo_vmware.api [None req-9e75f160-ec7f-4a3b-9503-b6838e2ef42a tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118103, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.708119] env[63531]: DEBUG oslo_vmware.api [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118101, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.192888} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.708606] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 893.709336] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26a6fba4-9702-4d0b-93ff-fcf4406d76a4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.732879] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Reconfiguring VM instance instance-0000004b to attach disk [datastore2] dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a/dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 893.733216] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c9d9ca45-6b9e-4d78-aa8f-d6e6d6339e8b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.754344] env[63531]: DEBUG oslo_vmware.api [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 893.754344] env[63531]: value = "task-1118104" [ 893.754344] env[63531]: _type = "Task" [ 893.754344] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.764347] env[63531]: DEBUG oslo_vmware.api [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118104, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.799309] env[63531]: DEBUG oslo_concurrency.lockutils [None req-eaa2182d-de39-454a-a00d-3595d2835100 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "a090135f-489d-457d-be36-ba6f61b71ab8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.562s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.805976] env[63531]: DEBUG oslo_concurrency.lockutils [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquiring lock "e95d3273-8216-47cc-95b6-99301366a827" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.806278] env[63531]: DEBUG oslo_concurrency.lockutils [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "e95d3273-8216-47cc-95b6-99301366a827" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.806603] env[63531]: DEBUG nova.compute.manager [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 893.807581] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b24f47a3-b658-4716-a1db-7c31252cc8fc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.813615] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8a9dc91a-31e6-4fb7-b47e-0b940db4e5dc tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.024s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.818501] env[63531]: DEBUG nova.compute.manager [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 893.824582] env[63531]: DEBUG oslo_concurrency.lockutils [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.176s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.824845] env[63531]: DEBUG nova.objects.instance [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Lazy-loading 'resources' on Instance uuid 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 893.826848] env[63531]: DEBUG nova.compute.manager [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63531) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 893.830590] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 893.830590] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ecc4b907-d348-4ddf-9781-fcb061001686 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.839961] env[63531]: DEBUG oslo_vmware.api [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 893.839961] env[63531]: value = "task-1118105" [ 893.839961] env[63531]: _type = "Task" [ 893.839961] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.867060] env[63531]: DEBUG oslo_vmware.api [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118105, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.868229] env[63531]: INFO nova.scheduler.client.report [None req-8a9dc91a-31e6-4fb7-b47e-0b940db4e5dc tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Deleted allocations for instance e7fb00a3-0d2e-4f54-950d-337307112d7a [ 893.871229] env[63531]: DEBUG nova.virt.hardware [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 893.871474] env[63531]: DEBUG nova.virt.hardware [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 893.871728] env[63531]: DEBUG nova.virt.hardware [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 893.871883] env[63531]: DEBUG nova.virt.hardware [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 893.871983] env[63531]: DEBUG nova.virt.hardware [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 893.872145] env[63531]: DEBUG nova.virt.hardware [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 893.872351] env[63531]: DEBUG nova.virt.hardware [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 893.872520] env[63531]: DEBUG nova.virt.hardware [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 893.872705] env[63531]: DEBUG nova.virt.hardware [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 893.872902] env[63531]: DEBUG nova.virt.hardware [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 893.873062] env[63531]: DEBUG nova.virt.hardware [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 893.874276] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85c94ada-3e5a-402b-9b6a-aa4f16c023e1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.882921] env[63531]: DEBUG oslo_vmware.api [None req-2b52d1c6-91c9-4f04-ad9c-74a98d502f8d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118102, 'name': PowerOffVM_Task, 'duration_secs': 0.33899} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.883912] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b52d1c6-91c9-4f04-ad9c-74a98d502f8d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 893.884134] env[63531]: DEBUG nova.compute.manager [None req-2b52d1c6-91c9-4f04-ad9c-74a98d502f8d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 893.885120] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb453132-401b-4cfc-91f9-e66e7646a012 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.892201] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55b87a73-b79a-470a-86e0-2e45abb6d4e4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.926616] env[63531]: DEBUG oslo_vmware.api [None req-9e75f160-ec7f-4a3b-9503-b6838e2ef42a tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118103, 'name': PowerOffVM_Task, 'duration_secs': 0.23308} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.926935] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e75f160-ec7f-4a3b-9503-b6838e2ef42a tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 893.938030] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-9e75f160-ec7f-4a3b-9503-b6838e2ef42a tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Reconfiguring VM instance instance-00000046 to detach disk 2001 {{(pid=63531) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 893.938030] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-51bd1d41-cbd7-4636-a3e7-f994f6bd1b2f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.961401] env[63531]: DEBUG oslo_vmware.api [None req-9e75f160-ec7f-4a3b-9503-b6838e2ef42a tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 893.961401] env[63531]: value = "task-1118106" [ 893.961401] env[63531]: _type = "Task" [ 893.961401] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.972134] env[63531]: DEBUG oslo_vmware.api [None req-9e75f160-ec7f-4a3b-9503-b6838e2ef42a tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118106, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.265629] env[63531]: DEBUG oslo_vmware.api [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118104, 'name': ReconfigVM_Task, 'duration_secs': 0.452857} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.266109] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Reconfigured VM instance instance-0000004b to attach disk [datastore2] dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a/dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 894.266830] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-853e47a8-643d-4598-9921-44ae7a963ec5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.274867] env[63531]: DEBUG oslo_vmware.api [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 894.274867] env[63531]: value = "task-1118107" [ 894.274867] env[63531]: _type = "Task" [ 894.274867] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.285465] env[63531]: DEBUG oslo_vmware.api [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118107, 'name': Rename_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.330560] env[63531]: DEBUG nova.objects.instance [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Lazy-loading 'numa_topology' on Instance uuid 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 894.332809] env[63531]: DEBUG nova.network.neutron [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] Successfully updated port: 161e87a1-1133-476f-9598-44c29595c36a {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 894.353100] env[63531]: DEBUG oslo_vmware.api [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118105, 'name': PowerOffVM_Task, 'duration_secs': 0.222169} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.353606] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 894.353916] env[63531]: DEBUG nova.compute.manager [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 894.355113] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ecad79e-5bef-485b-9736-276978032de9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.390785] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8a9dc91a-31e6-4fb7-b47e-0b940db4e5dc tempest-ServerRescueTestJSON-232073217 tempest-ServerRescueTestJSON-232073217-project-member] Lock "e7fb00a3-0d2e-4f54-950d-337307112d7a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.209s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.410120] env[63531]: DEBUG oslo_concurrency.lockutils [None req-2b52d1c6-91c9-4f04-ad9c-74a98d502f8d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "9a3fad46-a15a-451c-bdab-a3c8cc8add07" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.593s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.425362] env[63531]: DEBUG nova.compute.manager [req-06164aff-38f5-4a9b-bc2d-2ddedacd8199 req-48daee82-06bb-477c-a4c8-a987ed251741 service nova] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] Received event network-vif-plugged-161e87a1-1133-476f-9598-44c29595c36a {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 894.425588] env[63531]: DEBUG oslo_concurrency.lockutils [req-06164aff-38f5-4a9b-bc2d-2ddedacd8199 req-48daee82-06bb-477c-a4c8-a987ed251741 service nova] Acquiring lock "5ad9fc53-104b-4a90-801a-bd1e8f12c1d7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.425806] env[63531]: DEBUG oslo_concurrency.lockutils [req-06164aff-38f5-4a9b-bc2d-2ddedacd8199 req-48daee82-06bb-477c-a4c8-a987ed251741 service nova] Lock "5ad9fc53-104b-4a90-801a-bd1e8f12c1d7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.425979] env[63531]: DEBUG oslo_concurrency.lockutils [req-06164aff-38f5-4a9b-bc2d-2ddedacd8199 req-48daee82-06bb-477c-a4c8-a987ed251741 service nova] Lock "5ad9fc53-104b-4a90-801a-bd1e8f12c1d7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.427969] env[63531]: DEBUG nova.compute.manager [req-06164aff-38f5-4a9b-bc2d-2ddedacd8199 req-48daee82-06bb-477c-a4c8-a987ed251741 service nova] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] No waiting events found dispatching network-vif-plugged-161e87a1-1133-476f-9598-44c29595c36a {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 894.427969] env[63531]: WARNING nova.compute.manager [req-06164aff-38f5-4a9b-bc2d-2ddedacd8199 req-48daee82-06bb-477c-a4c8-a987ed251741 service nova] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] Received unexpected event network-vif-plugged-161e87a1-1133-476f-9598-44c29595c36a for instance with vm_state building and task_state spawning. [ 894.474812] env[63531]: DEBUG oslo_vmware.api [None req-9e75f160-ec7f-4a3b-9503-b6838e2ef42a tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118106, 'name': ReconfigVM_Task, 'duration_secs': 0.329923} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.475197] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-9e75f160-ec7f-4a3b-9503-b6838e2ef42a tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Reconfigured VM instance instance-00000046 to detach disk 2001 {{(pid=63531) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 894.475406] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e75f160-ec7f-4a3b-9503-b6838e2ef42a tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 894.475874] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3be2bbbd-f131-40e3-bfc5-477c69ffc61a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.486092] env[63531]: DEBUG oslo_vmware.api [None req-9e75f160-ec7f-4a3b-9503-b6838e2ef42a tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 894.486092] env[63531]: value = "task-1118108" [ 894.486092] env[63531]: _type = "Task" [ 894.486092] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.497178] env[63531]: DEBUG oslo_vmware.api [None req-9e75f160-ec7f-4a3b-9503-b6838e2ef42a tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118108, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.514292] env[63531]: DEBUG nova.compute.manager [req-8f0b444f-eaba-49de-b25c-ff2ebbf1e6bf req-66680b17-6b64-4e8d-b880-34c8064567e5 service nova] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Received event network-changed-17da826c-920c-4044-93cb-be8c8cf01c13 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 894.514452] env[63531]: DEBUG nova.compute.manager [req-8f0b444f-eaba-49de-b25c-ff2ebbf1e6bf req-66680b17-6b64-4e8d-b880-34c8064567e5 service nova] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Refreshing instance network info cache due to event network-changed-17da826c-920c-4044-93cb-be8c8cf01c13. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 894.514723] env[63531]: DEBUG oslo_concurrency.lockutils [req-8f0b444f-eaba-49de-b25c-ff2ebbf1e6bf req-66680b17-6b64-4e8d-b880-34c8064567e5 service nova] Acquiring lock "refresh_cache-a090135f-489d-457d-be36-ba6f61b71ab8" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 894.514919] env[63531]: DEBUG oslo_concurrency.lockutils [req-8f0b444f-eaba-49de-b25c-ff2ebbf1e6bf req-66680b17-6b64-4e8d-b880-34c8064567e5 service nova] Acquired lock "refresh_cache-a090135f-489d-457d-be36-ba6f61b71ab8" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.515090] env[63531]: DEBUG nova.network.neutron [req-8f0b444f-eaba-49de-b25c-ff2ebbf1e6bf req-66680b17-6b64-4e8d-b880-34c8064567e5 service nova] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Refreshing network info cache for port 17da826c-920c-4044-93cb-be8c8cf01c13 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 894.791380] env[63531]: DEBUG oslo_vmware.api [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118107, 'name': Rename_Task, 'duration_secs': 0.245592} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.791819] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 894.792038] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-092b0971-3522-436d-968d-3f3de7119197 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.802650] env[63531]: DEBUG oslo_vmware.api [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 894.802650] env[63531]: value = "task-1118109" [ 894.802650] env[63531]: _type = "Task" [ 894.802650] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.813071] env[63531]: DEBUG oslo_vmware.api [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118109, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.837792] env[63531]: DEBUG nova.objects.base [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Object Instance<2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f> lazy-loaded attributes: resources,numa_topology {{(pid=63531) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 894.843720] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "refresh_cache-5ad9fc53-104b-4a90-801a-bd1e8f12c1d7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 894.843720] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquired lock "refresh_cache-5ad9fc53-104b-4a90-801a-bd1e8f12c1d7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.843720] env[63531]: DEBUG nova.network.neutron [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 894.873182] env[63531]: DEBUG oslo_concurrency.lockutils [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "e95d3273-8216-47cc-95b6-99301366a827" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.067s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.999906] env[63531]: DEBUG oslo_vmware.api [None req-9e75f160-ec7f-4a3b-9503-b6838e2ef42a tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118108, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.206891] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eef01adc-a4f8-4bdd-b798-8a64539d1e7e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.216176] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e9a4037-f4d1-4640-b7ef-4987639d8cfd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.249430] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f1ac48f-e391-4b6d-89cc-2ef30158309a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.259533] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5ec8701-ec4e-48db-9cef-f04b9e84df0d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.277156] env[63531]: DEBUG nova.compute.provider_tree [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 895.313202] env[63531]: DEBUG oslo_vmware.api [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118109, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.384705] env[63531]: DEBUG oslo_concurrency.lockutils [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.454417] env[63531]: DEBUG nova.network.neutron [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 895.502323] env[63531]: DEBUG oslo_vmware.api [None req-9e75f160-ec7f-4a3b-9503-b6838e2ef42a tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118108, 'name': PowerOnVM_Task, 'duration_secs': 0.547336} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.502633] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e75f160-ec7f-4a3b-9503-b6838e2ef42a tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 895.502884] env[63531]: DEBUG nova.compute.manager [None req-9e75f160-ec7f-4a3b-9503-b6838e2ef42a tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 895.503785] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f729f944-c2fd-4f61-b186-ebb490f212f5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.611634] env[63531]: DEBUG nova.network.neutron [req-8f0b444f-eaba-49de-b25c-ff2ebbf1e6bf req-66680b17-6b64-4e8d-b880-34c8064567e5 service nova] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Updated VIF entry in instance network info cache for port 17da826c-920c-4044-93cb-be8c8cf01c13. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 895.612039] env[63531]: DEBUG nova.network.neutron [req-8f0b444f-eaba-49de-b25c-ff2ebbf1e6bf req-66680b17-6b64-4e8d-b880-34c8064567e5 service nova] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Updating instance_info_cache with network_info: [{"id": "17da826c-920c-4044-93cb-be8c8cf01c13", "address": "fa:16:3e:9a:ac:ad", "network": {"id": "caf86428-59a5-4e50-9135-2ab8473af64d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1673994696-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.214", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ba8b284854242c392aec5326e996239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap17da826c-92", "ovs_interfaceid": "17da826c-920c-4044-93cb-be8c8cf01c13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 895.664867] env[63531]: DEBUG nova.network.neutron [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] Updating instance_info_cache with network_info: [{"id": "161e87a1-1133-476f-9598-44c29595c36a", "address": "fa:16:3e:f3:d5:51", "network": {"id": "80641cc0-680b-40ff-8d1b-f5a5f736a88f", "bridge": "br-int", "label": "tempest-ServersTestJSON-666288463-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "63900312a30b4942854f552758c039cf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "26472e27-9835-4f87-ab7f-ca24dfee4e83", "external-id": "nsx-vlan-transportzone-335", "segmentation_id": 335, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap161e87a1-11", "ovs_interfaceid": "161e87a1-1133-476f-9598-44c29595c36a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 895.783812] env[63531]: DEBUG nova.scheduler.client.report [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 895.817531] env[63531]: DEBUG oslo_vmware.api [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118109, 'name': PowerOnVM_Task, 'duration_secs': 0.788164} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.817531] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 895.817531] env[63531]: INFO nova.compute.manager [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Took 8.57 seconds to spawn the instance on the hypervisor. [ 895.817531] env[63531]: DEBUG nova.compute.manager [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 895.818041] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-650fd649-4b64-4b4a-ad16-cec767134760 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.114605] env[63531]: DEBUG oslo_concurrency.lockutils [req-8f0b444f-eaba-49de-b25c-ff2ebbf1e6bf req-66680b17-6b64-4e8d-b880-34c8064567e5 service nova] Releasing lock "refresh_cache-a090135f-489d-457d-be36-ba6f61b71ab8" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.169017] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Releasing lock "refresh_cache-5ad9fc53-104b-4a90-801a-bd1e8f12c1d7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.169017] env[63531]: DEBUG nova.compute.manager [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] Instance network_info: |[{"id": "161e87a1-1133-476f-9598-44c29595c36a", "address": "fa:16:3e:f3:d5:51", "network": {"id": "80641cc0-680b-40ff-8d1b-f5a5f736a88f", "bridge": "br-int", "label": "tempest-ServersTestJSON-666288463-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "63900312a30b4942854f552758c039cf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "26472e27-9835-4f87-ab7f-ca24dfee4e83", "external-id": "nsx-vlan-transportzone-335", "segmentation_id": 335, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap161e87a1-11", "ovs_interfaceid": "161e87a1-1133-476f-9598-44c29595c36a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 896.169017] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f3:d5:51', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '26472e27-9835-4f87-ab7f-ca24dfee4e83', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '161e87a1-1133-476f-9598-44c29595c36a', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 896.176616] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Creating folder: Project (63900312a30b4942854f552758c039cf). Parent ref: group-v244585. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 896.177596] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-368fa7f3-5f4d-4ca3-ad46-f0201ae5d353 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.193734] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Created folder: Project (63900312a30b4942854f552758c039cf) in parent group-v244585. [ 896.193734] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Creating folder: Instances. Parent ref: group-v244736. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 896.193734] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-841cac3f-8850-4682-9a1e-4ee8d645e70b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.203084] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Created folder: Instances in parent group-v244736. [ 896.203524] env[63531]: DEBUG oslo.service.loopingcall [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 896.203860] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 896.204359] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-06435eee-7b00-4f99-9e81-78dcff5f3426 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.229133] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 896.229133] env[63531]: value = "task-1118112" [ 896.229133] env[63531]: _type = "Task" [ 896.229133] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.236449] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118112, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.291516] env[63531]: DEBUG oslo_concurrency.lockutils [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.465s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.293423] env[63531]: DEBUG oslo_concurrency.lockutils [None req-29b04bcd-cd4b-4a53-9b9f-4a106df45c8e tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.488s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.296123] env[63531]: DEBUG nova.objects.instance [None req-29b04bcd-cd4b-4a53-9b9f-4a106df45c8e tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Lazy-loading 'resources' on Instance uuid 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 896.348024] env[63531]: INFO nova.compute.manager [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Took 24.10 seconds to build instance. [ 896.453979] env[63531]: DEBUG nova.compute.manager [req-082af1ad-f3bb-482f-9ba2-7f3d2d845db4 req-ef58162b-b227-4bfc-9780-45063217fad8 service nova] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] Received event network-changed-161e87a1-1133-476f-9598-44c29595c36a {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 896.454246] env[63531]: DEBUG nova.compute.manager [req-082af1ad-f3bb-482f-9ba2-7f3d2d845db4 req-ef58162b-b227-4bfc-9780-45063217fad8 service nova] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] Refreshing instance network info cache due to event network-changed-161e87a1-1133-476f-9598-44c29595c36a. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 896.454465] env[63531]: DEBUG oslo_concurrency.lockutils [req-082af1ad-f3bb-482f-9ba2-7f3d2d845db4 req-ef58162b-b227-4bfc-9780-45063217fad8 service nova] Acquiring lock "refresh_cache-5ad9fc53-104b-4a90-801a-bd1e8f12c1d7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.454622] env[63531]: DEBUG oslo_concurrency.lockutils [req-082af1ad-f3bb-482f-9ba2-7f3d2d845db4 req-ef58162b-b227-4bfc-9780-45063217fad8 service nova] Acquired lock "refresh_cache-5ad9fc53-104b-4a90-801a-bd1e8f12c1d7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.454820] env[63531]: DEBUG nova.network.neutron [req-082af1ad-f3bb-482f-9ba2-7f3d2d845db4 req-ef58162b-b227-4bfc-9780-45063217fad8 service nova] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] Refreshing network info cache for port 161e87a1-1133-476f-9598-44c29595c36a {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 896.739849] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118112, 'name': CreateVM_Task, 'duration_secs': 0.494178} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.740057] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 896.741086] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.741086] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.741297] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 896.741883] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dd2a5614-d71a-4ef1-abbc-1024c84e72ad {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.747746] env[63531]: DEBUG oslo_vmware.api [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 896.747746] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]526f4344-a334-1a3d-47d0-489ed6d806b4" [ 896.747746] env[63531]: _type = "Task" [ 896.747746] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.757131] env[63531]: DEBUG oslo_vmware.api [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]526f4344-a334-1a3d-47d0-489ed6d806b4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.806329] env[63531]: DEBUG oslo_concurrency.lockutils [None req-80aab6cc-8b6c-4507-bdfd-005d82506c27 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Lock "2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 43.499s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.806329] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Lock "2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 19.237s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.806329] env[63531]: INFO nova.compute.manager [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Unshelving [ 896.847246] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f46774df-6630-4244-a34d-e7348317b5f7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.616s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.883595] env[63531]: DEBUG oslo_concurrency.lockutils [None req-935264d7-e19f-4031-a0a0-6f80fb185255 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquiring lock "e95d3273-8216-47cc-95b6-99301366a827" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.887131] env[63531]: DEBUG oslo_concurrency.lockutils [None req-935264d7-e19f-4031-a0a0-6f80fb185255 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "e95d3273-8216-47cc-95b6-99301366a827" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.887131] env[63531]: DEBUG oslo_concurrency.lockutils [None req-935264d7-e19f-4031-a0a0-6f80fb185255 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquiring lock "e95d3273-8216-47cc-95b6-99301366a827-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.887131] env[63531]: DEBUG oslo_concurrency.lockutils [None req-935264d7-e19f-4031-a0a0-6f80fb185255 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "e95d3273-8216-47cc-95b6-99301366a827-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.887131] env[63531]: DEBUG oslo_concurrency.lockutils [None req-935264d7-e19f-4031-a0a0-6f80fb185255 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "e95d3273-8216-47cc-95b6-99301366a827-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.896936] env[63531]: INFO nova.compute.manager [None req-935264d7-e19f-4031-a0a0-6f80fb185255 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Terminating instance [ 896.906371] env[63531]: DEBUG nova.compute.manager [None req-935264d7-e19f-4031-a0a0-6f80fb185255 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 896.906371] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-935264d7-e19f-4031-a0a0-6f80fb185255 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 896.906371] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d56421cf-d948-4fa3-afc3-4d8de85494fe {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.916280] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-935264d7-e19f-4031-a0a0-6f80fb185255 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 896.919749] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9d5b165e-82fb-45cb-8d05-e4380dc15da0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.991293] env[63531]: DEBUG nova.compute.manager [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Stashing vm_state: stopped {{(pid=63531) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 897.014062] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-935264d7-e19f-4031-a0a0-6f80fb185255 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 897.014473] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-935264d7-e19f-4031-a0a0-6f80fb185255 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 897.014581] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-935264d7-e19f-4031-a0a0-6f80fb185255 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Deleting the datastore file [datastore1] e95d3273-8216-47cc-95b6-99301366a827 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 897.014821] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8437f46e-e4df-48b2-ae8f-aa3dd4182808 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.025035] env[63531]: DEBUG oslo_vmware.api [None req-935264d7-e19f-4031-a0a0-6f80fb185255 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 897.025035] env[63531]: value = "task-1118114" [ 897.025035] env[63531]: _type = "Task" [ 897.025035] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.041293] env[63531]: DEBUG oslo_vmware.api [None req-935264d7-e19f-4031-a0a0-6f80fb185255 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118114, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.204103] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24d39077-7953-4b1f-9888-c2da1561668e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.213334] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc23eb49-2bf6-405c-b889-1f6772c05bdb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.251323] env[63531]: DEBUG nova.network.neutron [req-082af1ad-f3bb-482f-9ba2-7f3d2d845db4 req-ef58162b-b227-4bfc-9780-45063217fad8 service nova] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] Updated VIF entry in instance network info cache for port 161e87a1-1133-476f-9598-44c29595c36a. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 897.252028] env[63531]: DEBUG nova.network.neutron [req-082af1ad-f3bb-482f-9ba2-7f3d2d845db4 req-ef58162b-b227-4bfc-9780-45063217fad8 service nova] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] Updating instance_info_cache with network_info: [{"id": "161e87a1-1133-476f-9598-44c29595c36a", "address": "fa:16:3e:f3:d5:51", "network": {"id": "80641cc0-680b-40ff-8d1b-f5a5f736a88f", "bridge": "br-int", "label": "tempest-ServersTestJSON-666288463-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "63900312a30b4942854f552758c039cf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "26472e27-9835-4f87-ab7f-ca24dfee4e83", "external-id": "nsx-vlan-transportzone-335", "segmentation_id": 335, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap161e87a1-11", "ovs_interfaceid": "161e87a1-1133-476f-9598-44c29595c36a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 897.259451] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46fe7cea-d082-4acd-80e4-5a51ae8ae019 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.266996] env[63531]: DEBUG oslo_vmware.api [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]526f4344-a334-1a3d-47d0-489ed6d806b4, 'name': SearchDatastore_Task, 'duration_secs': 0.014158} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.269558] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.269849] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 897.270188] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.270349] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.270554] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 897.271516] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d6e64f6c-2c55-4241-ace4-cbc33a48c937 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.274641] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2237f476-3d64-4fde-8f9f-200ca9d77c0b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.290974] env[63531]: DEBUG nova.compute.provider_tree [None req-29b04bcd-cd4b-4a53-9b9f-4a106df45c8e tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 897.295815] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 897.295815] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 897.295815] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-52aaaf6c-3bc2-4b2b-963e-2a583eba42d9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.301482] env[63531]: DEBUG oslo_vmware.api [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 897.301482] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52b33386-d38f-9f52-b142-14f80e318872" [ 897.301482] env[63531]: _type = "Task" [ 897.301482] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.313438] env[63531]: DEBUG oslo_vmware.api [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52b33386-d38f-9f52-b142-14f80e318872, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.447508] env[63531]: DEBUG oslo_vmware.rw_handles [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f97289-5f27-b8d5-8c36-8b0cbd3d7139/disk-0.vmdk. {{(pid=63531) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 897.448473] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86c5b3f1-0c71-4262-b554-a9765a7f41d7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.455371] env[63531]: DEBUG oslo_vmware.rw_handles [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f97289-5f27-b8d5-8c36-8b0cbd3d7139/disk-0.vmdk is in state: ready. {{(pid=63531) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 897.455556] env[63531]: ERROR oslo_vmware.rw_handles [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f97289-5f27-b8d5-8c36-8b0cbd3d7139/disk-0.vmdk due to incomplete transfer. [ 897.455838] env[63531]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-c8a84f5f-9240-4863-9b44-6605623dbf91 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.463765] env[63531]: DEBUG oslo_vmware.rw_handles [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f97289-5f27-b8d5-8c36-8b0cbd3d7139/disk-0.vmdk. {{(pid=63531) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 897.463988] env[63531]: DEBUG nova.virt.vmwareapi.images [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Uploaded image d2fca167-8767-4b16-b64a-61fd11e58a0e to the Glance image server {{(pid=63531) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 897.466104] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Destroying the VM {{(pid=63531) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 897.466374] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-5297478d-c9ef-40cd-8882-b4f99500b86f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.473907] env[63531]: DEBUG oslo_vmware.api [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Waiting for the task: (returnval){ [ 897.473907] env[63531]: value = "task-1118115" [ 897.473907] env[63531]: _type = "Task" [ 897.473907] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.483587] env[63531]: DEBUG oslo_vmware.api [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1118115, 'name': Destroy_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.515011] env[63531]: DEBUG oslo_concurrency.lockutils [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.536766] env[63531]: DEBUG oslo_vmware.api [None req-935264d7-e19f-4031-a0a0-6f80fb185255 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118114, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.24722} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.537084] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-935264d7-e19f-4031-a0a0-6f80fb185255 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 897.537671] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-935264d7-e19f-4031-a0a0-6f80fb185255 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 897.537876] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-935264d7-e19f-4031-a0a0-6f80fb185255 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 897.538156] env[63531]: INFO nova.compute.manager [None req-935264d7-e19f-4031-a0a0-6f80fb185255 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Took 0.63 seconds to destroy the instance on the hypervisor. [ 897.538438] env[63531]: DEBUG oslo.service.loopingcall [None req-935264d7-e19f-4031-a0a0-6f80fb185255 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 897.538660] env[63531]: DEBUG nova.compute.manager [-] [instance: e95d3273-8216-47cc-95b6-99301366a827] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 897.538776] env[63531]: DEBUG nova.network.neutron [-] [instance: e95d3273-8216-47cc-95b6-99301366a827] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 897.761418] env[63531]: DEBUG oslo_concurrency.lockutils [req-082af1ad-f3bb-482f-9ba2-7f3d2d845db4 req-ef58162b-b227-4bfc-9780-45063217fad8 service nova] Releasing lock "refresh_cache-5ad9fc53-104b-4a90-801a-bd1e8f12c1d7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.796253] env[63531]: DEBUG nova.scheduler.client.report [None req-29b04bcd-cd4b-4a53-9b9f-4a106df45c8e tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 897.813896] env[63531]: DEBUG oslo_vmware.api [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52b33386-d38f-9f52-b142-14f80e318872, 'name': SearchDatastore_Task, 'duration_secs': 0.013212} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.814763] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8b355f74-71f7-4ef4-bd5c-afa00b448a15 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.825306] env[63531]: DEBUG oslo_vmware.api [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 897.825306] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52244e88-c930-b317-3ede-b0b550301f18" [ 897.825306] env[63531]: _type = "Task" [ 897.825306] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.830944] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.835231] env[63531]: DEBUG oslo_vmware.api [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52244e88-c930-b317-3ede-b0b550301f18, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.990020] env[63531]: DEBUG oslo_vmware.api [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1118115, 'name': Destroy_Task} progress is 33%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.093470] env[63531]: DEBUG oslo_concurrency.lockutils [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Acquiring lock "e065f052-ba3e-4783-8953-5dc200d1f3e9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.093470] env[63531]: DEBUG oslo_concurrency.lockutils [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Lock "e065f052-ba3e-4783-8953-5dc200d1f3e9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.125374] env[63531]: INFO nova.compute.manager [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Rebuilding instance [ 898.176218] env[63531]: DEBUG nova.compute.manager [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 898.177157] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e402d596-15ab-4dec-8456-bc15f9debbfb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.304381] env[63531]: DEBUG oslo_concurrency.lockutils [None req-29b04bcd-cd4b-4a53-9b9f-4a106df45c8e tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.011s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.306445] env[63531]: DEBUG oslo_concurrency.lockutils [None req-31ea6a4d-757c-44c3-8294-8cbb5267d005 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.041s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.307179] env[63531]: DEBUG nova.objects.instance [None req-31ea6a4d-757c-44c3-8294-8cbb5267d005 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Lazy-loading 'resources' on Instance uuid 3eeaa6f9-60d9-43fe-a438-fbf2ceececae {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 898.330458] env[63531]: INFO nova.scheduler.client.report [None req-29b04bcd-cd4b-4a53-9b9f-4a106df45c8e tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Deleted allocations for instance 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f [ 898.341655] env[63531]: DEBUG oslo_vmware.api [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52244e88-c930-b317-3ede-b0b550301f18, 'name': SearchDatastore_Task, 'duration_secs': 0.03181} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.343593] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.343841] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7/5ad9fc53-104b-4a90-801a-bd1e8f12c1d7.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 898.344124] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-89ce4f10-cf2a-4abc-8837-a96023cd55a2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.353238] env[63531]: DEBUG oslo_vmware.api [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 898.353238] env[63531]: value = "task-1118116" [ 898.353238] env[63531]: _type = "Task" [ 898.353238] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.359769] env[63531]: DEBUG nova.network.neutron [-] [instance: e95d3273-8216-47cc-95b6-99301366a827] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 898.366114] env[63531]: DEBUG oslo_vmware.api [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118116, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.485360] env[63531]: DEBUG nova.compute.manager [req-2969d957-291c-468a-adab-7562e5bc55c7 req-369f2f6c-dffc-476e-a4c6-bf0f2d813ac5 service nova] [instance: e95d3273-8216-47cc-95b6-99301366a827] Received event network-vif-deleted-a806c2a6-5a84-4114-aedb-8bb2da53a147 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 898.490301] env[63531]: DEBUG oslo_vmware.api [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1118115, 'name': Destroy_Task} progress is 100%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.595778] env[63531]: DEBUG nova.compute.manager [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 898.689783] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 898.690134] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8867a6a6-f4c7-4f4f-9ade-dc71c25ebb61 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.700924] env[63531]: DEBUG oslo_vmware.api [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 898.700924] env[63531]: value = "task-1118117" [ 898.700924] env[63531]: _type = "Task" [ 898.700924] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.712776] env[63531]: DEBUG oslo_vmware.api [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118117, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.839407] env[63531]: DEBUG oslo_concurrency.lockutils [None req-29b04bcd-cd4b-4a53-9b9f-4a106df45c8e tempest-ListImageFiltersTestJSON-797817625 tempest-ListImageFiltersTestJSON-797817625-project-member] Lock "2998fb3d-6677-4fb3-952a-c74d8c6b7b9f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.044s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.868047] env[63531]: INFO nova.compute.manager [-] [instance: e95d3273-8216-47cc-95b6-99301366a827] Took 1.33 seconds to deallocate network for instance. [ 898.868047] env[63531]: DEBUG oslo_vmware.api [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118116, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.986227] env[63531]: DEBUG oslo_vmware.api [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1118115, 'name': Destroy_Task} progress is 100%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.118715] env[63531]: DEBUG oslo_concurrency.lockutils [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.131014] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc381b49-cd7c-4788-ae77-c66c2cf7760b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.139056] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93b063b9-6e52-4381-8540-1d54a9c99204 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.169670] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71d11b08-9d66-4bdc-acf2-ea7b6a6dee34 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.178050] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e123d17-62af-4dad-b7c1-bb1e676976ec {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.192758] env[63531]: DEBUG nova.compute.provider_tree [None req-31ea6a4d-757c-44c3-8294-8cbb5267d005 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 899.212971] env[63531]: DEBUG oslo_vmware.api [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118117, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.365964] env[63531]: DEBUG oslo_vmware.api [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118116, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.679451} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.366055] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7/5ad9fc53-104b-4a90-801a-bd1e8f12c1d7.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 899.366223] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 899.366491] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5d1b2c86-876e-4318-a46c-f5d56f5aa080 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.374444] env[63531]: DEBUG oslo_vmware.api [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 899.374444] env[63531]: value = "task-1118118" [ 899.374444] env[63531]: _type = "Task" [ 899.374444] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.380266] env[63531]: DEBUG oslo_concurrency.lockutils [None req-935264d7-e19f-4031-a0a0-6f80fb185255 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.384714] env[63531]: DEBUG oslo_vmware.api [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118118, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.487641] env[63531]: DEBUG oslo_vmware.api [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1118115, 'name': Destroy_Task, 'duration_secs': 1.8973} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.487931] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Destroyed the VM [ 899.488188] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Deleting Snapshot of the VM instance {{(pid=63531) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 899.488520] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-eb3c5a9c-ff25-4fa3-ad7a-eb32c0034b6f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.495444] env[63531]: DEBUG oslo_vmware.api [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Waiting for the task: (returnval){ [ 899.495444] env[63531]: value = "task-1118119" [ 899.495444] env[63531]: _type = "Task" [ 899.495444] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.503686] env[63531]: DEBUG oslo_vmware.api [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1118119, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.696044] env[63531]: DEBUG nova.scheduler.client.report [None req-31ea6a4d-757c-44c3-8294-8cbb5267d005 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 899.712570] env[63531]: DEBUG oslo_vmware.api [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118117, 'name': PowerOffVM_Task, 'duration_secs': 0.962462} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.713563] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 899.713811] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 899.714599] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9716e8af-1a56-469f-8eee-7ebd4a48f7cc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.722748] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 899.723550] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d0001299-b453-4cc6-b7e9-8e0f12c38d7c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.816186] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 899.816473] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Deleting contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 899.816550] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Deleting the datastore file [datastore2] dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 899.816846] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-407f47ae-4ba6-4e08-b15f-dacfcbe95bd9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.823869] env[63531]: DEBUG oslo_vmware.api [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 899.823869] env[63531]: value = "task-1118121" [ 899.823869] env[63531]: _type = "Task" [ 899.823869] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.833088] env[63531]: DEBUG oslo_vmware.api [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118121, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.843612] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquiring lock "ccf00b09-29fe-4a76-a8af-97774acd77de" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.843802] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Lock "ccf00b09-29fe-4a76-a8af-97774acd77de" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.885419] env[63531]: DEBUG oslo_vmware.api [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118118, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07485} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.885717] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 899.886598] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3abc7bf4-8cf9-436b-befa-b5b170ddfacf {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.911636] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] Reconfiguring VM instance instance-0000004c to attach disk [datastore2] 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7/5ad9fc53-104b-4a90-801a-bd1e8f12c1d7.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 899.912729] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ccd46773-4114-4d26-91d5-f178104aebd3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.939378] env[63531]: DEBUG oslo_vmware.api [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 899.939378] env[63531]: value = "task-1118122" [ 899.939378] env[63531]: _type = "Task" [ 899.939378] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.948577] env[63531]: DEBUG oslo_vmware.api [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118122, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.009683] env[63531]: DEBUG oslo_vmware.api [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1118119, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.200793] env[63531]: DEBUG oslo_concurrency.lockutils [None req-31ea6a4d-757c-44c3-8294-8cbb5267d005 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.894s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.203480] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5cd60179-ca4c-47ad-8ed1-0bb74872da0a tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.348s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.203740] env[63531]: DEBUG nova.objects.instance [None req-5cd60179-ca4c-47ad-8ed1-0bb74872da0a tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Lazy-loading 'resources' on Instance uuid 45b05c54-6865-47a1-ba93-90ad3e1ba07e {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 900.221040] env[63531]: INFO nova.scheduler.client.report [None req-31ea6a4d-757c-44c3-8294-8cbb5267d005 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Deleted allocations for instance 3eeaa6f9-60d9-43fe-a438-fbf2ceececae [ 900.334952] env[63531]: DEBUG oslo_vmware.api [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118121, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.234388} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.335206] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 900.335362] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Deleted contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 900.335706] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 900.346109] env[63531]: DEBUG nova.compute.manager [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 900.449634] env[63531]: DEBUG oslo_vmware.api [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118122, 'name': ReconfigVM_Task, 'duration_secs': 0.360058} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.449844] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] Reconfigured VM instance instance-0000004c to attach disk [datastore2] 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7/5ad9fc53-104b-4a90-801a-bd1e8f12c1d7.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 900.450538] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a1c80784-53a0-4ece-904b-964800672566 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.457977] env[63531]: DEBUG oslo_vmware.api [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 900.457977] env[63531]: value = "task-1118123" [ 900.457977] env[63531]: _type = "Task" [ 900.457977] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.466209] env[63531]: DEBUG oslo_vmware.api [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118123, 'name': Rename_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.506192] env[63531]: DEBUG oslo_vmware.api [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1118119, 'name': RemoveSnapshot_Task, 'duration_secs': 0.827573} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.506635] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Deleted Snapshot of the VM instance {{(pid=63531) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 900.506962] env[63531]: DEBUG nova.compute.manager [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 900.507816] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a645abe4-569d-4aca-91a3-16726e716a6a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.652628] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 900.652913] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 900.729488] env[63531]: DEBUG oslo_concurrency.lockutils [None req-31ea6a4d-757c-44c3-8294-8cbb5267d005 tempest-ServerShowV247Test-2044195129 tempest-ServerShowV247Test-2044195129-project-member] Lock "3eeaa6f9-60d9-43fe-a438-fbf2ceececae" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.329s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.876865] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.971247] env[63531]: DEBUG oslo_vmware.api [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118123, 'name': Rename_Task, 'duration_secs': 0.344173} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.974729] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 900.975378] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-74036c2b-2d50-499f-ae29-47f76acee5e1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.984873] env[63531]: DEBUG oslo_vmware.api [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 900.984873] env[63531]: value = "task-1118124" [ 900.984873] env[63531]: _type = "Task" [ 900.984873] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.998579] env[63531]: DEBUG oslo_vmware.api [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118124, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.024872] env[63531]: INFO nova.compute.manager [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Shelve offloading [ 901.027125] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 901.027468] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2c4aceb7-6eda-4516-8021-fe8314a7e56e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.030974] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abffc7ee-3dea-4d04-83d6-2514549d5b54 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.043838] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4275d7cb-5a83-4fdc-a596-8dbb50a3df3f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.047927] env[63531]: DEBUG oslo_vmware.api [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Waiting for the task: (returnval){ [ 901.047927] env[63531]: value = "task-1118125" [ 901.047927] env[63531]: _type = "Task" [ 901.047927] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.077845] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afe58d9e-c875-4978-b78b-646fdc050402 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.084285] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] VM already powered off {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 901.084508] env[63531]: DEBUG nova.compute.manager [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 901.085259] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-499e4ddf-4d93-4427-984d-fecca6f6a97f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.091815] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3591b4d7-6307-4d29-ad36-5473383ec5c0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.097059] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Acquiring lock "refresh_cache-1ca7b851-2a06-4181-8271-58aafcd322d6" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.097269] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Acquired lock "refresh_cache-1ca7b851-2a06-4181-8271-58aafcd322d6" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.097471] env[63531]: DEBUG nova.network.neutron [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 901.108832] env[63531]: DEBUG nova.compute.provider_tree [None req-5cd60179-ca4c-47ad-8ed1-0bb74872da0a tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 901.161524] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 901.161524] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Starting heal instance info cache {{(pid=63531) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 901.374308] env[63531]: DEBUG nova.virt.hardware [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 901.374583] env[63531]: DEBUG nova.virt.hardware [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 901.374583] env[63531]: DEBUG nova.virt.hardware [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 901.374774] env[63531]: DEBUG nova.virt.hardware [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 901.374925] env[63531]: DEBUG nova.virt.hardware [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 901.375103] env[63531]: DEBUG nova.virt.hardware [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 901.375342] env[63531]: DEBUG nova.virt.hardware [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 901.375565] env[63531]: DEBUG nova.virt.hardware [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 901.375752] env[63531]: DEBUG nova.virt.hardware [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 901.375927] env[63531]: DEBUG nova.virt.hardware [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 901.376123] env[63531]: DEBUG nova.virt.hardware [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 901.377015] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c09169e7-e912-4446-82d4-c06116198e55 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.388836] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b8c8e0d-cbcf-4cf5-ba5c-0924addebbe8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.406751] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a5:e0:6e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bc6e6fe1-c4f5-4389-a49f-0978060eebb4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3e1c2a04-40a1-4e53-8770-dd615bd761eb', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 901.414752] env[63531]: DEBUG oslo.service.loopingcall [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 901.415011] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 901.415280] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8f47b17f-bc27-44da-9f67-6fc6d7dda152 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.438124] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 901.438124] env[63531]: value = "task-1118126" [ 901.438124] env[63531]: _type = "Task" [ 901.438124] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.447948] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118126, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.495270] env[63531]: DEBUG oslo_vmware.api [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118124, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.614044] env[63531]: DEBUG nova.scheduler.client.report [None req-5cd60179-ca4c-47ad-8ed1-0bb74872da0a tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 901.937969] env[63531]: DEBUG nova.network.neutron [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Updating instance_info_cache with network_info: [{"id": "a70530fb-7805-40a0-bea5-a2a666dc679d", "address": "fa:16:3e:78:c3:90", "network": {"id": "73e6d3ed-2016-498a-bda6-fd7dfbff93da", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1501868918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f727ddb992dc4c06954f25f6a1613b4a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c20f5114-0866-45b3-9a7c-62f113ff83fa", "external-id": "nsx-vlan-transportzone-47", "segmentation_id": 47, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa70530fb-78", "ovs_interfaceid": "a70530fb-7805-40a0-bea5-a2a666dc679d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.950219] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118126, 'name': CreateVM_Task} progress is 25%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.995584] env[63531]: DEBUG oslo_vmware.api [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118124, 'name': PowerOnVM_Task, 'duration_secs': 0.513424} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.999021] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 901.999021] env[63531]: INFO nova.compute.manager [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] Took 8.18 seconds to spawn the instance on the hypervisor. [ 901.999021] env[63531]: DEBUG nova.compute.manager [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 901.999021] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de6f1b73-5eb0-40a8-98bb-850bd024b376 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.119755] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5cd60179-ca4c-47ad-8ed1-0bb74872da0a tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.914s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.120421] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.587s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.122386] env[63531]: INFO nova.compute.claims [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 902.149203] env[63531]: INFO nova.scheduler.client.report [None req-5cd60179-ca4c-47ad-8ed1-0bb74872da0a tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Deleted allocations for instance 45b05c54-6865-47a1-ba93-90ad3e1ba07e [ 902.440800] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Releasing lock "refresh_cache-1ca7b851-2a06-4181-8271-58aafcd322d6" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.458018] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118126, 'name': CreateVM_Task, 'duration_secs': 0.642268} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.458018] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 902.458018] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.458018] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.459400] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 902.459729] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8532c38d-1067-437b-b878-a26d7eba9c8c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.465821] env[63531]: DEBUG oslo_vmware.api [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 902.465821] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]527246e4-fef0-2fde-df25-b60a119cff23" [ 902.465821] env[63531]: _type = "Task" [ 902.465821] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.474998] env[63531]: DEBUG oslo_vmware.api [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]527246e4-fef0-2fde-df25-b60a119cff23, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.516760] env[63531]: INFO nova.compute.manager [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] Took 28.98 seconds to build instance. [ 902.656805] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5cd60179-ca4c-47ad-8ed1-0bb74872da0a tempest-ServersTestFqdnHostnames-805218362 tempest-ServersTestFqdnHostnames-805218362-project-member] Lock "45b05c54-6865-47a1-ba93-90ad3e1ba07e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.206s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.783163] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 902.784160] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-049113dd-7d7b-47d0-a512-6fb67bfc1182 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.792119] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 902.792371] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2bcc411c-f8e8-46e2-8097-ffe03203e8c1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.863170] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 902.863518] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Deleting contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 902.863615] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Deleting the datastore file [datastore2] 1ca7b851-2a06-4181-8271-58aafcd322d6 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 902.863968] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b50542d5-7d01-411b-9fed-4eef5034d041 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.873023] env[63531]: DEBUG oslo_vmware.api [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Waiting for the task: (returnval){ [ 902.873023] env[63531]: value = "task-1118128" [ 902.873023] env[63531]: _type = "Task" [ 902.873023] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.882255] env[63531]: DEBUG oslo_vmware.api [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1118128, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.922961] env[63531]: DEBUG nova.compute.manager [req-ee9d4679-d405-44e4-a5a2-66599227c642 req-128c206b-3b53-4d30-bd68-70fdeeec61c9 service nova] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Received event network-vif-unplugged-a70530fb-7805-40a0-bea5-a2a666dc679d {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 902.923840] env[63531]: DEBUG oslo_concurrency.lockutils [req-ee9d4679-d405-44e4-a5a2-66599227c642 req-128c206b-3b53-4d30-bd68-70fdeeec61c9 service nova] Acquiring lock "1ca7b851-2a06-4181-8271-58aafcd322d6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.926686] env[63531]: DEBUG oslo_concurrency.lockutils [req-ee9d4679-d405-44e4-a5a2-66599227c642 req-128c206b-3b53-4d30-bd68-70fdeeec61c9 service nova] Lock "1ca7b851-2a06-4181-8271-58aafcd322d6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.926686] env[63531]: DEBUG oslo_concurrency.lockutils [req-ee9d4679-d405-44e4-a5a2-66599227c642 req-128c206b-3b53-4d30-bd68-70fdeeec61c9 service nova] Lock "1ca7b851-2a06-4181-8271-58aafcd322d6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.926686] env[63531]: DEBUG nova.compute.manager [req-ee9d4679-d405-44e4-a5a2-66599227c642 req-128c206b-3b53-4d30-bd68-70fdeeec61c9 service nova] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] No waiting events found dispatching network-vif-unplugged-a70530fb-7805-40a0-bea5-a2a666dc679d {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 902.926686] env[63531]: WARNING nova.compute.manager [req-ee9d4679-d405-44e4-a5a2-66599227c642 req-128c206b-3b53-4d30-bd68-70fdeeec61c9 service nova] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Received unexpected event network-vif-unplugged-a70530fb-7805-40a0-bea5-a2a666dc679d for instance with vm_state shelved and task_state shelving_offloading. [ 902.978176] env[63531]: DEBUG oslo_vmware.api [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]527246e4-fef0-2fde-df25-b60a119cff23, 'name': SearchDatastore_Task, 'duration_secs': 0.010432} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.978176] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.978333] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 902.978697] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.980098] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.980098] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 902.980098] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f381ba54-1eb2-4cbd-a0e9-b58465a1cde5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.997532] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 902.998046] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 903.000096] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fbcbaaa0-1235-4da5-978b-f63d6796f4ce {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.008177] env[63531]: DEBUG oslo_vmware.api [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 903.008177] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]525fb97e-485e-0832-0437-e16f12b05d27" [ 903.008177] env[63531]: _type = "Task" [ 903.008177] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.019439] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8ea1c7ee-7bf1-46b9-a840-b9be50c90e83 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "5ad9fc53-104b-4a90-801a-bd1e8f12c1d7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.490s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.020157] env[63531]: DEBUG oslo_vmware.api [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]525fb97e-485e-0832-0437-e16f12b05d27, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.389445] env[63531]: DEBUG oslo_vmware.api [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1118128, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.147966} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.389568] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 903.389698] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Deleted contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 903.389935] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 903.418285] env[63531]: INFO nova.scheduler.client.report [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Deleted allocations for instance 1ca7b851-2a06-4181-8271-58aafcd322d6 [ 903.456024] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2853e25-9bc8-4899-b937-8df8a90aa566 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.462989] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dc08a88-1c8c-490f-8b3f-ecdc2138654d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.494755] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66f6708c-ce81-4e61-bc1f-383af7abbb1e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.503363] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25b21d31-bce9-4009-93f1-703bd5816b16 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.522565] env[63531]: DEBUG nova.compute.provider_tree [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 903.527121] env[63531]: DEBUG oslo_vmware.api [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]525fb97e-485e-0832-0437-e16f12b05d27, 'name': SearchDatastore_Task, 'duration_secs': 0.015517} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.528179] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ec70e0c2-2e25-48d9-baee-1e8fe3236775 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.534020] env[63531]: DEBUG oslo_vmware.api [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 903.534020] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52e86c1f-33bd-30cd-e13c-66305c7e5889" [ 903.534020] env[63531]: _type = "Task" [ 903.534020] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.543553] env[63531]: DEBUG oslo_vmware.api [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52e86c1f-33bd-30cd-e13c-66305c7e5889, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.675375] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Acquiring lock "refresh_cache-1ca7b851-2a06-4181-8271-58aafcd322d6" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 903.675533] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Acquired lock "refresh_cache-1ca7b851-2a06-4181-8271-58aafcd322d6" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.675700] env[63531]: DEBUG nova.network.neutron [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Forcefully refreshing network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 903.925173] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.029595] env[63531]: DEBUG nova.scheduler.client.report [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 904.046612] env[63531]: DEBUG oslo_vmware.api [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52e86c1f-33bd-30cd-e13c-66305c7e5889, 'name': SearchDatastore_Task, 'duration_secs': 0.010182} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.046612] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 904.046612] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a/dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 904.046948] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-602fe35e-af82-4692-8777-d74f2c7dd381 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.056164] env[63531]: DEBUG oslo_vmware.api [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 904.056164] env[63531]: value = "task-1118129" [ 904.056164] env[63531]: _type = "Task" [ 904.056164] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.065494] env[63531]: DEBUG oslo_vmware.api [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118129, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.536238] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.416s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.536772] env[63531]: DEBUG nova.compute.manager [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 904.545185] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.425s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.545185] env[63531]: INFO nova.compute.claims [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 06ec91b2-14cb-4db8-8fab-e6c4183b1116] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 904.572810] env[63531]: DEBUG oslo_vmware.api [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118129, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.967331] env[63531]: DEBUG nova.network.neutron [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Updating instance_info_cache with network_info: [{"id": "a70530fb-7805-40a0-bea5-a2a666dc679d", "address": "fa:16:3e:78:c3:90", "network": {"id": "73e6d3ed-2016-498a-bda6-fd7dfbff93da", "bridge": null, "label": "tempest-ServersNegativeTestJSON-1501868918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f727ddb992dc4c06954f25f6a1613b4a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapa70530fb-78", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.047118] env[63531]: DEBUG nova.compute.utils [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 905.047118] env[63531]: DEBUG nova.compute.manager [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 905.047771] env[63531]: DEBUG nova.network.neutron [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 905.075285] env[63531]: DEBUG oslo_vmware.api [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118129, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.094236] env[63531]: DEBUG nova.policy [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '597d424341124b7db23dc7a104107148', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '308eabafea9c4188a58a0f1c22074d2f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 905.113971] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "3fb0499d-8074-4e63-be9f-380730416cc4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.114233] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "3fb0499d-8074-4e63-be9f-380730416cc4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.269022] env[63531]: DEBUG nova.compute.manager [req-1c636e1b-edcf-41b7-8e60-c54034432ae9 req-91d1964b-b359-4cee-909d-643bda04221c service nova] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Received event network-changed-a70530fb-7805-40a0-bea5-a2a666dc679d {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 905.269022] env[63531]: DEBUG nova.compute.manager [req-1c636e1b-edcf-41b7-8e60-c54034432ae9 req-91d1964b-b359-4cee-909d-643bda04221c service nova] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Refreshing instance network info cache due to event network-changed-a70530fb-7805-40a0-bea5-a2a666dc679d. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 905.269022] env[63531]: DEBUG oslo_concurrency.lockutils [req-1c636e1b-edcf-41b7-8e60-c54034432ae9 req-91d1964b-b359-4cee-909d-643bda04221c service nova] Acquiring lock "refresh_cache-1ca7b851-2a06-4181-8271-58aafcd322d6" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.377354] env[63531]: DEBUG nova.network.neutron [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Successfully created port: 0997f575-14bc-4813-9df6-b18e4c1c2186 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 905.469874] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Releasing lock "refresh_cache-1ca7b851-2a06-4181-8271-58aafcd322d6" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.470087] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Updated the network info_cache for instance {{(pid=63531) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 905.470384] env[63531]: DEBUG oslo_concurrency.lockutils [req-1c636e1b-edcf-41b7-8e60-c54034432ae9 req-91d1964b-b359-4cee-909d-643bda04221c service nova] Acquired lock "refresh_cache-1ca7b851-2a06-4181-8271-58aafcd322d6" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.470570] env[63531]: DEBUG nova.network.neutron [req-1c636e1b-edcf-41b7-8e60-c54034432ae9 req-91d1964b-b359-4cee-909d-643bda04221c service nova] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Refreshing network info cache for port a70530fb-7805-40a0-bea5-a2a666dc679d {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 905.472504] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 905.472807] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 905.472978] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 905.474897] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 905.475135] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 905.475412] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 905.475498] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63531) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 905.475655] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager.update_available_resource {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 905.552581] env[63531]: DEBUG nova.compute.manager [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 905.582960] env[63531]: DEBUG oslo_vmware.api [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118129, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.480324} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.583455] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a/dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 905.583570] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 905.583828] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d1cc2189-0fdc-4a59-8dd4-ff4c4dc38973 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.594960] env[63531]: DEBUG oslo_vmware.api [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 905.594960] env[63531]: value = "task-1118130" [ 905.594960] env[63531]: _type = "Task" [ 905.594960] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.611914] env[63531]: DEBUG oslo_vmware.api [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118130, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.618598] env[63531]: DEBUG nova.compute.manager [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3fb0499d-8074-4e63-be9f-380730416cc4] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 905.930482] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc141e29-8d02-41f9-a65e-50a256eb800f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.940380] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ef7ea5a-4893-442c-bba6-6eeba8dfad57 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.988400] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.989358] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5d31005-029d-4513-a9a9-8bbd32100d76 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.998758] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d4e8e1a-1a39-4d97-b956-9d4134d57205 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.013420] env[63531]: DEBUG nova.compute.provider_tree [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 906.108844] env[63531]: DEBUG oslo_vmware.api [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118130, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.290311} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.109151] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 906.110094] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c5285fc-701c-4203-9410-b2a67166d8c6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.151163] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Reconfiguring VM instance instance-0000004b to attach disk [datastore1] dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a/dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 906.157572] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5728a485-a93b-43d4-b9b1-2d03013e770b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.180242] env[63531]: DEBUG oslo_vmware.api [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 906.180242] env[63531]: value = "task-1118131" [ 906.180242] env[63531]: _type = "Task" [ 906.180242] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.190746] env[63531]: DEBUG oslo_vmware.api [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118131, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.191830] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.243877] env[63531]: DEBUG nova.network.neutron [req-1c636e1b-edcf-41b7-8e60-c54034432ae9 req-91d1964b-b359-4cee-909d-643bda04221c service nova] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Updated VIF entry in instance network info cache for port a70530fb-7805-40a0-bea5-a2a666dc679d. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 906.244359] env[63531]: DEBUG nova.network.neutron [req-1c636e1b-edcf-41b7-8e60-c54034432ae9 req-91d1964b-b359-4cee-909d-643bda04221c service nova] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Updating instance_info_cache with network_info: [{"id": "a70530fb-7805-40a0-bea5-a2a666dc679d", "address": "fa:16:3e:78:c3:90", "network": {"id": "73e6d3ed-2016-498a-bda6-fd7dfbff93da", "bridge": null, "label": "tempest-ServersNegativeTestJSON-1501868918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f727ddb992dc4c06954f25f6a1613b4a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapa70530fb-78", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 906.516996] env[63531]: DEBUG nova.scheduler.client.report [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 906.545108] env[63531]: DEBUG oslo_concurrency.lockutils [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Acquiring lock "1ca7b851-2a06-4181-8271-58aafcd322d6" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.569928] env[63531]: DEBUG nova.compute.manager [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 906.599314] env[63531]: DEBUG nova.virt.hardware [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 906.599878] env[63531]: DEBUG nova.virt.hardware [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 906.600245] env[63531]: DEBUG nova.virt.hardware [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 906.600650] env[63531]: DEBUG nova.virt.hardware [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 906.600993] env[63531]: DEBUG nova.virt.hardware [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 906.601355] env[63531]: DEBUG nova.virt.hardware [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 906.601775] env[63531]: DEBUG nova.virt.hardware [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 906.602157] env[63531]: DEBUG nova.virt.hardware [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 906.602542] env[63531]: DEBUG nova.virt.hardware [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 906.602908] env[63531]: DEBUG nova.virt.hardware [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 906.603292] env[63531]: DEBUG nova.virt.hardware [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 906.604939] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1b134da-5ed2-4f8e-b26b-a226bd48c4f5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.616680] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-604822c0-33de-4773-b9a2-7be5f8f93bb5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.692278] env[63531]: DEBUG oslo_vmware.api [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118131, 'name': ReconfigVM_Task, 'duration_secs': 0.292547} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.692596] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Reconfigured VM instance instance-0000004b to attach disk [datastore1] dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a/dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 906.693338] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d348dd38-b5ea-4485-89bf-95aeef55a725 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.701629] env[63531]: DEBUG oslo_vmware.api [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 906.701629] env[63531]: value = "task-1118132" [ 906.701629] env[63531]: _type = "Task" [ 906.701629] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.710911] env[63531]: DEBUG oslo_vmware.api [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118132, 'name': Rename_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.747739] env[63531]: DEBUG oslo_concurrency.lockutils [req-1c636e1b-edcf-41b7-8e60-c54034432ae9 req-91d1964b-b359-4cee-909d-643bda04221c service nova] Releasing lock "refresh_cache-1ca7b851-2a06-4181-8271-58aafcd322d6" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.023489] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.480s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.024068] env[63531]: DEBUG nova.compute.manager [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 06ec91b2-14cb-4db8-8fab-e6c4183b1116] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 907.029727] env[63531]: DEBUG oslo_concurrency.lockutils [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.813s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.030717] env[63531]: INFO nova.compute.claims [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 907.205184] env[63531]: DEBUG nova.network.neutron [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Successfully updated port: 0997f575-14bc-4813-9df6-b18e4c1c2186 {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 907.222709] env[63531]: DEBUG oslo_vmware.api [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118132, 'name': Rename_Task, 'duration_secs': 0.183707} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.223445] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 907.224641] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3460d352-f341-4ac5-abbf-12569f6b6073 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.233579] env[63531]: DEBUG oslo_vmware.api [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 907.233579] env[63531]: value = "task-1118133" [ 907.233579] env[63531]: _type = "Task" [ 907.233579] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.244812] env[63531]: DEBUG oslo_vmware.api [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118133, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.298151] env[63531]: DEBUG nova.compute.manager [req-34d7e497-9fe4-4203-bc8e-c95228efea3b req-f7d0e49d-9e6a-4b0b-a864-314355e7ac2e service nova] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Received event network-vif-plugged-0997f575-14bc-4813-9df6-b18e4c1c2186 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 907.298151] env[63531]: DEBUG oslo_concurrency.lockutils [req-34d7e497-9fe4-4203-bc8e-c95228efea3b req-f7d0e49d-9e6a-4b0b-a864-314355e7ac2e service nova] Acquiring lock "6118e242-c0db-4707-a5f9-3d12b823935f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.298151] env[63531]: DEBUG oslo_concurrency.lockutils [req-34d7e497-9fe4-4203-bc8e-c95228efea3b req-f7d0e49d-9e6a-4b0b-a864-314355e7ac2e service nova] Lock "6118e242-c0db-4707-a5f9-3d12b823935f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.298777] env[63531]: DEBUG oslo_concurrency.lockutils [req-34d7e497-9fe4-4203-bc8e-c95228efea3b req-f7d0e49d-9e6a-4b0b-a864-314355e7ac2e service nova] Lock "6118e242-c0db-4707-a5f9-3d12b823935f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.299132] env[63531]: DEBUG nova.compute.manager [req-34d7e497-9fe4-4203-bc8e-c95228efea3b req-f7d0e49d-9e6a-4b0b-a864-314355e7ac2e service nova] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] No waiting events found dispatching network-vif-plugged-0997f575-14bc-4813-9df6-b18e4c1c2186 {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 907.299463] env[63531]: WARNING nova.compute.manager [req-34d7e497-9fe4-4203-bc8e-c95228efea3b req-f7d0e49d-9e6a-4b0b-a864-314355e7ac2e service nova] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Received unexpected event network-vif-plugged-0997f575-14bc-4813-9df6-b18e4c1c2186 for instance with vm_state building and task_state spawning. [ 907.299739] env[63531]: DEBUG nova.compute.manager [req-34d7e497-9fe4-4203-bc8e-c95228efea3b req-f7d0e49d-9e6a-4b0b-a864-314355e7ac2e service nova] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Received event network-changed-0997f575-14bc-4813-9df6-b18e4c1c2186 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 907.300135] env[63531]: DEBUG nova.compute.manager [req-34d7e497-9fe4-4203-bc8e-c95228efea3b req-f7d0e49d-9e6a-4b0b-a864-314355e7ac2e service nova] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Refreshing instance network info cache due to event network-changed-0997f575-14bc-4813-9df6-b18e4c1c2186. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 907.300625] env[63531]: DEBUG oslo_concurrency.lockutils [req-34d7e497-9fe4-4203-bc8e-c95228efea3b req-f7d0e49d-9e6a-4b0b-a864-314355e7ac2e service nova] Acquiring lock "refresh_cache-6118e242-c0db-4707-a5f9-3d12b823935f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.300924] env[63531]: DEBUG oslo_concurrency.lockutils [req-34d7e497-9fe4-4203-bc8e-c95228efea3b req-f7d0e49d-9e6a-4b0b-a864-314355e7ac2e service nova] Acquired lock "refresh_cache-6118e242-c0db-4707-a5f9-3d12b823935f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.306019] env[63531]: DEBUG nova.network.neutron [req-34d7e497-9fe4-4203-bc8e-c95228efea3b req-f7d0e49d-9e6a-4b0b-a864-314355e7ac2e service nova] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Refreshing network info cache for port 0997f575-14bc-4813-9df6-b18e4c1c2186 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 907.539102] env[63531]: DEBUG nova.compute.utils [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 907.542532] env[63531]: DEBUG nova.compute.manager [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 06ec91b2-14cb-4db8-8fab-e6c4183b1116] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 907.542532] env[63531]: DEBUG nova.network.neutron [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 06ec91b2-14cb-4db8-8fab-e6c4183b1116] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 907.600829] env[63531]: DEBUG nova.policy [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0372c22d1e9e4f66a0417643dab4a2ec', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '94a95667c00947ea9183a6307c569c90', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 907.705938] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "refresh_cache-6118e242-c0db-4707-a5f9-3d12b823935f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.746167] env[63531]: DEBUG oslo_vmware.api [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118133, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.860454] env[63531]: DEBUG nova.network.neutron [req-34d7e497-9fe4-4203-bc8e-c95228efea3b req-f7d0e49d-9e6a-4b0b-a864-314355e7ac2e service nova] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 908.015287] env[63531]: DEBUG nova.network.neutron [req-34d7e497-9fe4-4203-bc8e-c95228efea3b req-f7d0e49d-9e6a-4b0b-a864-314355e7ac2e service nova] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 908.048637] env[63531]: DEBUG nova.compute.manager [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 06ec91b2-14cb-4db8-8fab-e6c4183b1116] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 908.064252] env[63531]: DEBUG nova.network.neutron [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 06ec91b2-14cb-4db8-8fab-e6c4183b1116] Successfully created port: 5c98d4e0-4d30-4104-8ee9-15361087529a {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 908.252204] env[63531]: DEBUG oslo_vmware.api [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118133, 'name': PowerOnVM_Task, 'duration_secs': 0.594447} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.253051] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 908.253051] env[63531]: DEBUG nova.compute.manager [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 908.256152] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6234b84a-611a-443a-a60e-51517fe50938 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.395512] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a898b2e8-6dc2-4799-bed1-4c878f168c6a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.403479] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b81cfb4e-1a65-4f40-a3c5-c99f4304ba02 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.436467] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fffe9e73-4d08-4e83-813e-06185cbd88bd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.444813] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d26901f4-6808-42fe-9a85-3540dab5b06d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.461548] env[63531]: DEBUG nova.compute.provider_tree [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 908.518740] env[63531]: DEBUG oslo_concurrency.lockutils [req-34d7e497-9fe4-4203-bc8e-c95228efea3b req-f7d0e49d-9e6a-4b0b-a864-314355e7ac2e service nova] Releasing lock "refresh_cache-6118e242-c0db-4707-a5f9-3d12b823935f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 908.519140] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquired lock "refresh_cache-6118e242-c0db-4707-a5f9-3d12b823935f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 908.520228] env[63531]: DEBUG nova.network.neutron [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 908.773990] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.965634] env[63531]: DEBUG nova.scheduler.client.report [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 909.053075] env[63531]: DEBUG nova.network.neutron [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 909.061132] env[63531]: DEBUG nova.compute.manager [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 06ec91b2-14cb-4db8-8fab-e6c4183b1116] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 909.104051] env[63531]: DEBUG nova.virt.hardware [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 909.104332] env[63531]: DEBUG nova.virt.hardware [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 909.104494] env[63531]: DEBUG nova.virt.hardware [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 909.104694] env[63531]: DEBUG nova.virt.hardware [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 909.104848] env[63531]: DEBUG nova.virt.hardware [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 909.105015] env[63531]: DEBUG nova.virt.hardware [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 909.105246] env[63531]: DEBUG nova.virt.hardware [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 909.105416] env[63531]: DEBUG nova.virt.hardware [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 909.105589] env[63531]: DEBUG nova.virt.hardware [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 909.105760] env[63531]: DEBUG nova.virt.hardware [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 909.105943] env[63531]: DEBUG nova.virt.hardware [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 909.106875] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-956a53dd-4823-4c9e-bd6b-17493a75ba6c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.117853] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6d04197-db97-4a34-83be-00ffa7b77156 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.216881] env[63531]: DEBUG nova.network.neutron [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Updating instance_info_cache with network_info: [{"id": "0997f575-14bc-4813-9df6-b18e4c1c2186", "address": "fa:16:3e:43:5d:33", "network": {"id": "0667c176-cfdf-4e43-b152-e00bdc224448", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-243715985-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "308eabafea9c4188a58a0f1c22074d2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8f580e6-1d86-41ee-9ebe-c531cb9299c6", "external-id": "nsx-vlan-transportzone-150", "segmentation_id": 150, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0997f575-14", "ovs_interfaceid": "0997f575-14bc-4813-9df6-b18e4c1c2186", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 909.474305] env[63531]: DEBUG oslo_concurrency.lockutils [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.443s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.474305] env[63531]: DEBUG nova.compute.manager [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 909.474504] env[63531]: DEBUG oslo_concurrency.lockutils [None req-817aa9b4-fcda-44b3-b34f-da4ac7528014 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.065s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.474765] env[63531]: DEBUG nova.objects.instance [None req-817aa9b4-fcda-44b3-b34f-da4ac7528014 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Lazy-loading 'resources' on Instance uuid 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 909.720941] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Releasing lock "refresh_cache-6118e242-c0db-4707-a5f9-3d12b823935f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.720941] env[63531]: DEBUG nova.compute.manager [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Instance network_info: |[{"id": "0997f575-14bc-4813-9df6-b18e4c1c2186", "address": "fa:16:3e:43:5d:33", "network": {"id": "0667c176-cfdf-4e43-b152-e00bdc224448", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-243715985-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "308eabafea9c4188a58a0f1c22074d2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8f580e6-1d86-41ee-9ebe-c531cb9299c6", "external-id": "nsx-vlan-transportzone-150", "segmentation_id": 150, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0997f575-14", "ovs_interfaceid": "0997f575-14bc-4813-9df6-b18e4c1c2186", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 909.721381] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:43:5d:33', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c8f580e6-1d86-41ee-9ebe-c531cb9299c6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0997f575-14bc-4813-9df6-b18e4c1c2186', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 909.730354] env[63531]: DEBUG oslo.service.loopingcall [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 909.730637] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 909.730878] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9c46237c-e608-4ae4-9d59-0f622a11db44 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.759449] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 909.759449] env[63531]: value = "task-1118134" [ 909.759449] env[63531]: _type = "Task" [ 909.759449] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.769352] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118134, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.770402] env[63531]: DEBUG nova.network.neutron [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 06ec91b2-14cb-4db8-8fab-e6c4183b1116] Successfully updated port: 5c98d4e0-4d30-4104-8ee9-15361087529a {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 909.928324] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9bd3dc4a-86f3-4a59-85fb-1184d578df53 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.928926] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9bd3dc4a-86f3-4a59-85fb-1184d578df53 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.929199] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9bd3dc4a-86f3-4a59-85fb-1184d578df53 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.929435] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9bd3dc4a-86f3-4a59-85fb-1184d578df53 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.929781] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9bd3dc4a-86f3-4a59-85fb-1184d578df53 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.936730] env[63531]: INFO nova.compute.manager [None req-9bd3dc4a-86f3-4a59-85fb-1184d578df53 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Terminating instance [ 909.941289] env[63531]: DEBUG nova.compute.manager [None req-9bd3dc4a-86f3-4a59-85fb-1184d578df53 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 909.941568] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-9bd3dc4a-86f3-4a59-85fb-1184d578df53 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 909.942565] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5813890-2ea0-474a-9c50-cab398447564 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.953190] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-9bd3dc4a-86f3-4a59-85fb-1184d578df53 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 909.953512] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6b811a89-2af1-4892-876e-44e0d015924d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.962202] env[63531]: DEBUG oslo_vmware.api [None req-9bd3dc4a-86f3-4a59-85fb-1184d578df53 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 909.962202] env[63531]: value = "task-1118135" [ 909.962202] env[63531]: _type = "Task" [ 909.962202] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.975358] env[63531]: DEBUG oslo_vmware.api [None req-9bd3dc4a-86f3-4a59-85fb-1184d578df53 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118135, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.978030] env[63531]: DEBUG nova.compute.utils [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 909.983124] env[63531]: DEBUG nova.compute.manager [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 909.983124] env[63531]: DEBUG nova.network.neutron [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 910.064631] env[63531]: DEBUG nova.compute.manager [req-4f732b67-4b35-4d2c-9a07-796f4c01ed3c req-339ab5d1-0a66-4781-89e8-288950edc292 service nova] [instance: 06ec91b2-14cb-4db8-8fab-e6c4183b1116] Received event network-vif-plugged-5c98d4e0-4d30-4104-8ee9-15361087529a {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 910.065077] env[63531]: DEBUG oslo_concurrency.lockutils [req-4f732b67-4b35-4d2c-9a07-796f4c01ed3c req-339ab5d1-0a66-4781-89e8-288950edc292 service nova] Acquiring lock "06ec91b2-14cb-4db8-8fab-e6c4183b1116-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.065141] env[63531]: DEBUG oslo_concurrency.lockutils [req-4f732b67-4b35-4d2c-9a07-796f4c01ed3c req-339ab5d1-0a66-4781-89e8-288950edc292 service nova] Lock "06ec91b2-14cb-4db8-8fab-e6c4183b1116-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.065300] env[63531]: DEBUG oslo_concurrency.lockutils [req-4f732b67-4b35-4d2c-9a07-796f4c01ed3c req-339ab5d1-0a66-4781-89e8-288950edc292 service nova] Lock "06ec91b2-14cb-4db8-8fab-e6c4183b1116-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.065597] env[63531]: DEBUG nova.compute.manager [req-4f732b67-4b35-4d2c-9a07-796f4c01ed3c req-339ab5d1-0a66-4781-89e8-288950edc292 service nova] [instance: 06ec91b2-14cb-4db8-8fab-e6c4183b1116] No waiting events found dispatching network-vif-plugged-5c98d4e0-4d30-4104-8ee9-15361087529a {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 910.065786] env[63531]: WARNING nova.compute.manager [req-4f732b67-4b35-4d2c-9a07-796f4c01ed3c req-339ab5d1-0a66-4781-89e8-288950edc292 service nova] [instance: 06ec91b2-14cb-4db8-8fab-e6c4183b1116] Received unexpected event network-vif-plugged-5c98d4e0-4d30-4104-8ee9-15361087529a for instance with vm_state building and task_state spawning. [ 910.078242] env[63531]: DEBUG nova.policy [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '30c2b29064b44c48ad7e76d445c65317', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '04a4c814245346a9b73253a25a822fef', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 910.271389] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118134, 'name': CreateVM_Task, 'duration_secs': 0.394895} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.274538] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 910.278929] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "refresh_cache-06ec91b2-14cb-4db8-8fab-e6c4183b1116" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.278929] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquired lock "refresh_cache-06ec91b2-14cb-4db8-8fab-e6c4183b1116" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.278929] env[63531]: DEBUG nova.network.neutron [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 06ec91b2-14cb-4db8-8fab-e6c4183b1116] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 910.278929] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.278929] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.279634] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 910.282694] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c4316db9-43ba-4955-a6d3-e34f84b27767 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.289624] env[63531]: DEBUG oslo_vmware.api [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 910.289624] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]529d6f9f-dced-20d8-6f50-98bb3cf758b1" [ 910.289624] env[63531]: _type = "Task" [ 910.289624] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.302362] env[63531]: DEBUG oslo_vmware.api [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]529d6f9f-dced-20d8-6f50-98bb3cf758b1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.468965] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3529ec8d-dc46-4777-92bb-1c4edd606f72 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.483027] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25d30438-815c-4138-a7ac-da766c46bef8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.485221] env[63531]: DEBUG oslo_vmware.api [None req-9bd3dc4a-86f3-4a59-85fb-1184d578df53 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118135, 'name': PowerOffVM_Task, 'duration_secs': 0.259604} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.485753] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-9bd3dc4a-86f3-4a59-85fb-1184d578df53 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 910.486111] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-9bd3dc4a-86f3-4a59-85fb-1184d578df53 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 910.487021] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-24ceda0a-e5ff-4cdb-bf60-7663e42b08db {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.489046] env[63531]: DEBUG nova.compute.manager [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 910.518453] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7f16c72-a54b-4025-a6e8-29a1fd80e16e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.530143] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f52bb9a6-c996-45d5-a92b-20cab4674ad8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.547410] env[63531]: DEBUG nova.compute.provider_tree [None req-817aa9b4-fcda-44b3-b34f-da4ac7528014 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 910.631180] env[63531]: DEBUG nova.network.neutron [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Successfully created port: d31275a8-8fba-48c3-be21-3b90d7cc05cf {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 910.643376] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-9bd3dc4a-86f3-4a59-85fb-1184d578df53 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 910.643640] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-9bd3dc4a-86f3-4a59-85fb-1184d578df53 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 910.644382] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-9bd3dc4a-86f3-4a59-85fb-1184d578df53 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Deleting the datastore file [datastore1] dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 910.644542] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-47bf38d7-9893-4189-a686-7c715a9af9cf {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.653713] env[63531]: DEBUG oslo_vmware.api [None req-9bd3dc4a-86f3-4a59-85fb-1184d578df53 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 910.653713] env[63531]: value = "task-1118137" [ 910.653713] env[63531]: _type = "Task" [ 910.653713] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.668607] env[63531]: DEBUG oslo_vmware.api [None req-9bd3dc4a-86f3-4a59-85fb-1184d578df53 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118137, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.744244] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "27c23b7c-a68f-43c9-a125-bd657feb3c5b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.744244] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "27c23b7c-a68f-43c9-a125-bd657feb3c5b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.803846] env[63531]: DEBUG oslo_vmware.api [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]529d6f9f-dced-20d8-6f50-98bb3cf758b1, 'name': SearchDatastore_Task, 'duration_secs': 0.029597} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.804344] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 910.805250] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 910.805682] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.805845] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.806057] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 910.806504] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-34161bb1-d839-4680-8380-fd8dbb807686 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.831628] env[63531]: DEBUG nova.network.neutron [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 06ec91b2-14cb-4db8-8fab-e6c4183b1116] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 910.841962] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 910.842610] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 910.843215] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d3748faa-9724-4940-9c44-a61858e57096 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.852446] env[63531]: DEBUG oslo_vmware.api [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 910.852446] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]520ec986-912b-fc4e-4b49-6151f6c53d26" [ 910.852446] env[63531]: _type = "Task" [ 910.852446] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.863095] env[63531]: DEBUG oslo_vmware.api [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]520ec986-912b-fc4e-4b49-6151f6c53d26, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.054108] env[63531]: DEBUG nova.scheduler.client.report [None req-817aa9b4-fcda-44b3-b34f-da4ac7528014 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 911.116224] env[63531]: DEBUG nova.network.neutron [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 06ec91b2-14cb-4db8-8fab-e6c4183b1116] Updating instance_info_cache with network_info: [{"id": "5c98d4e0-4d30-4104-8ee9-15361087529a", "address": "fa:16:3e:0f:59:0f", "network": {"id": "ea2d47d3-a502-4afd-9bc5-ee4a37017901", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1361282648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94a95667c00947ea9183a6307c569c90", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6c45fd-e930-495a-9cb7-df84eda443b1", "external-id": "nsx-vlan-transportzone-407", "segmentation_id": 407, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c98d4e0-4d", "ovs_interfaceid": "5c98d4e0-4d30-4104-8ee9-15361087529a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 911.172195] env[63531]: DEBUG oslo_vmware.api [None req-9bd3dc4a-86f3-4a59-85fb-1184d578df53 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118137, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.310644} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.172195] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-9bd3dc4a-86f3-4a59-85fb-1184d578df53 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 911.172195] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-9bd3dc4a-86f3-4a59-85fb-1184d578df53 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 911.172368] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-9bd3dc4a-86f3-4a59-85fb-1184d578df53 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 911.172434] env[63531]: INFO nova.compute.manager [None req-9bd3dc4a-86f3-4a59-85fb-1184d578df53 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Took 1.23 seconds to destroy the instance on the hypervisor. [ 911.172683] env[63531]: DEBUG oslo.service.loopingcall [None req-9bd3dc4a-86f3-4a59-85fb-1184d578df53 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 911.172976] env[63531]: DEBUG nova.compute.manager [-] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 911.173167] env[63531]: DEBUG nova.network.neutron [-] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 911.247750] env[63531]: DEBUG nova.compute.manager [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 911.365108] env[63531]: DEBUG oslo_vmware.api [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]520ec986-912b-fc4e-4b49-6151f6c53d26, 'name': SearchDatastore_Task, 'duration_secs': 0.022275} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.366232] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-beb44ca2-a1c5-4a5b-932c-ce5cb6cb20a3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.372541] env[63531]: DEBUG oslo_vmware.api [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 911.372541] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52da5f18-b532-4eeb-8164-c24362113dce" [ 911.372541] env[63531]: _type = "Task" [ 911.372541] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.381687] env[63531]: DEBUG oslo_vmware.api [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52da5f18-b532-4eeb-8164-c24362113dce, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.524361] env[63531]: DEBUG nova.compute.manager [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 911.549856] env[63531]: DEBUG nova.virt.hardware [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 911.550347] env[63531]: DEBUG nova.virt.hardware [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 911.550636] env[63531]: DEBUG nova.virt.hardware [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 911.550850] env[63531]: DEBUG nova.virt.hardware [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 911.551976] env[63531]: DEBUG nova.virt.hardware [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 911.552211] env[63531]: DEBUG nova.virt.hardware [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 911.552548] env[63531]: DEBUG nova.virt.hardware [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 911.552695] env[63531]: DEBUG nova.virt.hardware [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 911.552908] env[63531]: DEBUG nova.virt.hardware [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 911.553129] env[63531]: DEBUG nova.virt.hardware [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 911.553350] env[63531]: DEBUG nova.virt.hardware [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 911.554413] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d249273-924a-4982-9459-4decc7339b92 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.558046] env[63531]: DEBUG oslo_concurrency.lockutils [None req-817aa9b4-fcda-44b3-b34f-da4ac7528014 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.084s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.560865] env[63531]: DEBUG oslo_concurrency.lockutils [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 16.178s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.561135] env[63531]: DEBUG nova.objects.instance [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: e95d3273-8216-47cc-95b6-99301366a827] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63531) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 911.570852] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f14f2a01-511d-425c-8f36-71278742e33a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.592979] env[63531]: INFO nova.scheduler.client.report [None req-817aa9b4-fcda-44b3-b34f-da4ac7528014 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Deleted allocations for instance 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47 [ 911.619201] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Releasing lock "refresh_cache-06ec91b2-14cb-4db8-8fab-e6c4183b1116" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.619560] env[63531]: DEBUG nova.compute.manager [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 06ec91b2-14cb-4db8-8fab-e6c4183b1116] Instance network_info: |[{"id": "5c98d4e0-4d30-4104-8ee9-15361087529a", "address": "fa:16:3e:0f:59:0f", "network": {"id": "ea2d47d3-a502-4afd-9bc5-ee4a37017901", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1361282648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94a95667c00947ea9183a6307c569c90", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6c45fd-e930-495a-9cb7-df84eda443b1", "external-id": "nsx-vlan-transportzone-407", "segmentation_id": 407, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c98d4e0-4d", "ovs_interfaceid": "5c98d4e0-4d30-4104-8ee9-15361087529a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 911.620285] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 06ec91b2-14cb-4db8-8fab-e6c4183b1116] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0f:59:0f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2b6c45fd-e930-495a-9cb7-df84eda443b1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5c98d4e0-4d30-4104-8ee9-15361087529a', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 911.629512] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Creating folder: Project (94a95667c00947ea9183a6307c569c90). Parent ref: group-v244585. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 911.630219] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-20a425a4-5e9f-428e-bc82-918ab868865b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.646342] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Created folder: Project (94a95667c00947ea9183a6307c569c90) in parent group-v244585. [ 911.646847] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Creating folder: Instances. Parent ref: group-v244741. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 911.647350] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6f121971-171d-4780-9e79-60f258b6eb02 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.658552] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Created folder: Instances in parent group-v244741. [ 911.659191] env[63531]: DEBUG oslo.service.loopingcall [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 911.659451] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 06ec91b2-14cb-4db8-8fab-e6c4183b1116] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 911.659725] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5f8ba002-c201-4ab5-bf0f-f12d5f3efbe4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.688958] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 911.688958] env[63531]: value = "task-1118140" [ 911.688958] env[63531]: _type = "Task" [ 911.688958] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.699833] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118140, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.767911] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.887024] env[63531]: DEBUG oslo_vmware.api [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52da5f18-b532-4eeb-8164-c24362113dce, 'name': SearchDatastore_Task, 'duration_secs': 0.011061} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.887024] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.887024] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 6118e242-c0db-4707-a5f9-3d12b823935f/6118e242-c0db-4707-a5f9-3d12b823935f.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 911.887024] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1a782be3-9f4f-4c3a-9876-6501b6512370 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.898099] env[63531]: DEBUG oslo_vmware.api [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 911.898099] env[63531]: value = "task-1118141" [ 911.898099] env[63531]: _type = "Task" [ 911.898099] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.916543] env[63531]: DEBUG oslo_vmware.api [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118141, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.949869] env[63531]: DEBUG nova.network.neutron [-] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.108248] env[63531]: DEBUG oslo_concurrency.lockutils [None req-817aa9b4-fcda-44b3-b34f-da4ac7528014 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Lock "1e5e28ef-38ce-47be-ac36-8bdf17ab0a47" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.661s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.170180] env[63531]: DEBUG nova.compute.manager [req-00d7e310-1ba9-444a-8939-7abd38c908e7 req-ce42205f-9515-45a4-be65-db57f3acce8b service nova] [instance: 06ec91b2-14cb-4db8-8fab-e6c4183b1116] Received event network-changed-5c98d4e0-4d30-4104-8ee9-15361087529a {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 912.170509] env[63531]: DEBUG nova.compute.manager [req-00d7e310-1ba9-444a-8939-7abd38c908e7 req-ce42205f-9515-45a4-be65-db57f3acce8b service nova] [instance: 06ec91b2-14cb-4db8-8fab-e6c4183b1116] Refreshing instance network info cache due to event network-changed-5c98d4e0-4d30-4104-8ee9-15361087529a. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 912.170839] env[63531]: DEBUG oslo_concurrency.lockutils [req-00d7e310-1ba9-444a-8939-7abd38c908e7 req-ce42205f-9515-45a4-be65-db57f3acce8b service nova] Acquiring lock "refresh_cache-06ec91b2-14cb-4db8-8fab-e6c4183b1116" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 912.171045] env[63531]: DEBUG oslo_concurrency.lockutils [req-00d7e310-1ba9-444a-8939-7abd38c908e7 req-ce42205f-9515-45a4-be65-db57f3acce8b service nova] Acquired lock "refresh_cache-06ec91b2-14cb-4db8-8fab-e6c4183b1116" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.174217] env[63531]: DEBUG nova.network.neutron [req-00d7e310-1ba9-444a-8939-7abd38c908e7 req-ce42205f-9515-45a4-be65-db57f3acce8b service nova] [instance: 06ec91b2-14cb-4db8-8fab-e6c4183b1116] Refreshing network info cache for port 5c98d4e0-4d30-4104-8ee9-15361087529a {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 912.204349] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118140, 'name': CreateVM_Task, 'duration_secs': 0.432531} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.204582] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 06ec91b2-14cb-4db8-8fab-e6c4183b1116] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 912.206205] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 912.206447] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.206899] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 912.209846] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-12376da7-7aa4-4197-bd7f-0d14443ba4d9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.216676] env[63531]: DEBUG oslo_vmware.api [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 912.216676] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52a8ed24-8cf0-7972-0e1f-d81da2af1498" [ 912.216676] env[63531]: _type = "Task" [ 912.216676] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.227465] env[63531]: DEBUG oslo_vmware.api [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52a8ed24-8cf0-7972-0e1f-d81da2af1498, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.413768] env[63531]: DEBUG oslo_vmware.api [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118141, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.454893] env[63531]: INFO nova.compute.manager [-] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Took 1.28 seconds to deallocate network for instance. [ 912.517838] env[63531]: DEBUG oslo_concurrency.lockutils [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Acquiring lock "55111b87-1e9a-4877-8689-987faa72a54e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.517838] env[63531]: DEBUG oslo_concurrency.lockutils [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Lock "55111b87-1e9a-4877-8689-987faa72a54e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.571981] env[63531]: DEBUG oslo_concurrency.lockutils [None req-2299f4f4-0e40-4bac-bf1e-02c1420c0b94 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.011s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.573184] env[63531]: DEBUG oslo_concurrency.lockutils [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 15.059s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.626946] env[63531]: DEBUG nova.network.neutron [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Successfully updated port: d31275a8-8fba-48c3-be21-3b90d7cc05cf {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 912.731402] env[63531]: DEBUG oslo_vmware.api [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52a8ed24-8cf0-7972-0e1f-d81da2af1498, 'name': SearchDatastore_Task, 'duration_secs': 0.059382} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.731719] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 912.731967] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 06ec91b2-14cb-4db8-8fab-e6c4183b1116] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 912.732236] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 912.732390] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.732580] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 912.732853] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2155127b-e671-4602-8acf-4b099e1d8e78 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.744156] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 912.744156] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 912.745207] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3cfe6feb-362e-4d1e-bf6e-7bf278aa20cd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.753746] env[63531]: DEBUG oslo_vmware.api [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 912.753746] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5254ac31-48e4-e0c5-c9fd-2ba639adf1c2" [ 912.753746] env[63531]: _type = "Task" [ 912.753746] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.767118] env[63531]: DEBUG oslo_vmware.api [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5254ac31-48e4-e0c5-c9fd-2ba639adf1c2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.911878] env[63531]: DEBUG oslo_vmware.api [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118141, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.521162} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.913796] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 6118e242-c0db-4707-a5f9-3d12b823935f/6118e242-c0db-4707-a5f9-3d12b823935f.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 912.914337] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 912.914825] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b4472d27-908d-4fc4-a07a-59c28e7dee1f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.921919] env[63531]: DEBUG nova.network.neutron [req-00d7e310-1ba9-444a-8939-7abd38c908e7 req-ce42205f-9515-45a4-be65-db57f3acce8b service nova] [instance: 06ec91b2-14cb-4db8-8fab-e6c4183b1116] Updated VIF entry in instance network info cache for port 5c98d4e0-4d30-4104-8ee9-15361087529a. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 912.922311] env[63531]: DEBUG nova.network.neutron [req-00d7e310-1ba9-444a-8939-7abd38c908e7 req-ce42205f-9515-45a4-be65-db57f3acce8b service nova] [instance: 06ec91b2-14cb-4db8-8fab-e6c4183b1116] Updating instance_info_cache with network_info: [{"id": "5c98d4e0-4d30-4104-8ee9-15361087529a", "address": "fa:16:3e:0f:59:0f", "network": {"id": "ea2d47d3-a502-4afd-9bc5-ee4a37017901", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1361282648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94a95667c00947ea9183a6307c569c90", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6c45fd-e930-495a-9cb7-df84eda443b1", "external-id": "nsx-vlan-transportzone-407", "segmentation_id": 407, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c98d4e0-4d", "ovs_interfaceid": "5c98d4e0-4d30-4104-8ee9-15361087529a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.935941] env[63531]: DEBUG oslo_vmware.api [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 912.935941] env[63531]: value = "task-1118142" [ 912.935941] env[63531]: _type = "Task" [ 912.935941] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.950277] env[63531]: DEBUG oslo_vmware.api [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118142, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.964829] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9bd3dc4a-86f3-4a59-85fb-1184d578df53 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.020338] env[63531]: DEBUG nova.compute.manager [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] [instance: 55111b87-1e9a-4877-8689-987faa72a54e] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 913.086371] env[63531]: INFO nova.compute.claims [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 913.129955] env[63531]: DEBUG oslo_concurrency.lockutils [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Acquiring lock "refresh_cache-4b5e5821-9def-4b53-be19-355d9e5f81ec" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.130246] env[63531]: DEBUG oslo_concurrency.lockutils [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Acquired lock "refresh_cache-4b5e5821-9def-4b53-be19-355d9e5f81ec" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.130289] env[63531]: DEBUG nova.network.neutron [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 913.273196] env[63531]: DEBUG oslo_vmware.api [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5254ac31-48e4-e0c5-c9fd-2ba639adf1c2, 'name': SearchDatastore_Task, 'duration_secs': 0.011116} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.273196] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-16c83c77-8a24-4f7a-9cdc-8a324e6d7934 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.279743] env[63531]: DEBUG oslo_vmware.api [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 913.279743] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52f33031-b0a5-bc3b-5e0d-181aeb98c714" [ 913.279743] env[63531]: _type = "Task" [ 913.279743] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.292945] env[63531]: DEBUG oslo_vmware.api [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52f33031-b0a5-bc3b-5e0d-181aeb98c714, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.427133] env[63531]: DEBUG oslo_concurrency.lockutils [req-00d7e310-1ba9-444a-8939-7abd38c908e7 req-ce42205f-9515-45a4-be65-db57f3acce8b service nova] Releasing lock "refresh_cache-06ec91b2-14cb-4db8-8fab-e6c4183b1116" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.428421] env[63531]: DEBUG nova.compute.manager [req-00d7e310-1ba9-444a-8939-7abd38c908e7 req-ce42205f-9515-45a4-be65-db57f3acce8b service nova] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Received event network-vif-deleted-3e1c2a04-40a1-4e53-8770-dd615bd761eb {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 913.447075] env[63531]: DEBUG oslo_vmware.api [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118142, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071164} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.447875] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 913.449282] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f25c7b52-ff2b-400d-8d34-4cf9e4436a8c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.474060] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Reconfiguring VM instance instance-0000004d to attach disk [datastore1] 6118e242-c0db-4707-a5f9-3d12b823935f/6118e242-c0db-4707-a5f9-3d12b823935f.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 913.474380] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-45097bd2-4a92-43eb-9810-c2be02b7902e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.496566] env[63531]: DEBUG oslo_vmware.api [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 913.496566] env[63531]: value = "task-1118143" [ 913.496566] env[63531]: _type = "Task" [ 913.496566] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.505112] env[63531]: DEBUG oslo_vmware.api [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118143, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.545837] env[63531]: DEBUG oslo_concurrency.lockutils [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.591334] env[63531]: INFO nova.compute.resource_tracker [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Updating resource usage from migration 72374b80-67c2-4204-8fdb-7c6eff83d384 [ 913.665071] env[63531]: DEBUG nova.network.neutron [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 913.794815] env[63531]: DEBUG oslo_vmware.api [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52f33031-b0a5-bc3b-5e0d-181aeb98c714, 'name': SearchDatastore_Task, 'duration_secs': 0.010645} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.794815] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.794815] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 06ec91b2-14cb-4db8-8fab-e6c4183b1116/06ec91b2-14cb-4db8-8fab-e6c4183b1116.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 913.795031] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d59efac3-d5aa-4a6a-bec6-0054c8010ba5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.803959] env[63531]: DEBUG oslo_vmware.api [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 913.803959] env[63531]: value = "task-1118144" [ 913.803959] env[63531]: _type = "Task" [ 913.803959] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.812204] env[63531]: DEBUG oslo_vmware.api [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118144, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.815563] env[63531]: DEBUG nova.network.neutron [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Updating instance_info_cache with network_info: [{"id": "d31275a8-8fba-48c3-be21-3b90d7cc05cf", "address": "fa:16:3e:b7:c3:f7", "network": {"id": "887fff70-ea19-458d-a945-e4f3ed6dc185", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2099566288-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "04a4c814245346a9b73253a25a822fef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49a4d142-3f97-47fe-b074-58923c46815e", "external-id": "nsx-vlan-transportzone-565", "segmentation_id": 565, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd31275a8-8f", "ovs_interfaceid": "d31275a8-8fba-48c3-be21-3b90d7cc05cf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 913.882644] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c771a5e5-f747-4b3f-9f4c-d9983e3850c8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.890886] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0df8c05-e352-4cc1-afaa-5b8b1e596ca8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.921451] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-832567ad-e133-4ca5-8989-4dd4d698e971 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.929612] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bb7b74b-a487-42f6-912c-7242191880e0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.945602] env[63531]: DEBUG nova.compute.provider_tree [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 914.009552] env[63531]: DEBUG oslo_vmware.api [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118143, 'name': ReconfigVM_Task, 'duration_secs': 0.277987} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.009906] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Reconfigured VM instance instance-0000004d to attach disk [datastore1] 6118e242-c0db-4707-a5f9-3d12b823935f/6118e242-c0db-4707-a5f9-3d12b823935f.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 914.010626] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-99feab31-1032-494a-94bb-52a494fd7ce6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.019728] env[63531]: DEBUG oslo_vmware.api [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 914.019728] env[63531]: value = "task-1118145" [ 914.019728] env[63531]: _type = "Task" [ 914.019728] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.030342] env[63531]: DEBUG oslo_vmware.api [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118145, 'name': Rename_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.199917] env[63531]: DEBUG nova.compute.manager [req-305bab81-7eb4-447d-ad2a-8e627c10b93c req-824cbf59-787d-417d-8208-5c9086fdba37 service nova] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Received event network-vif-plugged-d31275a8-8fba-48c3-be21-3b90d7cc05cf {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 914.200205] env[63531]: DEBUG oslo_concurrency.lockutils [req-305bab81-7eb4-447d-ad2a-8e627c10b93c req-824cbf59-787d-417d-8208-5c9086fdba37 service nova] Acquiring lock "4b5e5821-9def-4b53-be19-355d9e5f81ec-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.200409] env[63531]: DEBUG oslo_concurrency.lockutils [req-305bab81-7eb4-447d-ad2a-8e627c10b93c req-824cbf59-787d-417d-8208-5c9086fdba37 service nova] Lock "4b5e5821-9def-4b53-be19-355d9e5f81ec-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.200594] env[63531]: DEBUG oslo_concurrency.lockutils [req-305bab81-7eb4-447d-ad2a-8e627c10b93c req-824cbf59-787d-417d-8208-5c9086fdba37 service nova] Lock "4b5e5821-9def-4b53-be19-355d9e5f81ec-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.200658] env[63531]: DEBUG nova.compute.manager [req-305bab81-7eb4-447d-ad2a-8e627c10b93c req-824cbf59-787d-417d-8208-5c9086fdba37 service nova] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] No waiting events found dispatching network-vif-plugged-d31275a8-8fba-48c3-be21-3b90d7cc05cf {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 914.200828] env[63531]: WARNING nova.compute.manager [req-305bab81-7eb4-447d-ad2a-8e627c10b93c req-824cbf59-787d-417d-8208-5c9086fdba37 service nova] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Received unexpected event network-vif-plugged-d31275a8-8fba-48c3-be21-3b90d7cc05cf for instance with vm_state building and task_state spawning. [ 914.200994] env[63531]: DEBUG nova.compute.manager [req-305bab81-7eb4-447d-ad2a-8e627c10b93c req-824cbf59-787d-417d-8208-5c9086fdba37 service nova] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Received event network-changed-d31275a8-8fba-48c3-be21-3b90d7cc05cf {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 914.201175] env[63531]: DEBUG nova.compute.manager [req-305bab81-7eb4-447d-ad2a-8e627c10b93c req-824cbf59-787d-417d-8208-5c9086fdba37 service nova] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Refreshing instance network info cache due to event network-changed-d31275a8-8fba-48c3-be21-3b90d7cc05cf. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 914.201348] env[63531]: DEBUG oslo_concurrency.lockutils [req-305bab81-7eb4-447d-ad2a-8e627c10b93c req-824cbf59-787d-417d-8208-5c9086fdba37 service nova] Acquiring lock "refresh_cache-4b5e5821-9def-4b53-be19-355d9e5f81ec" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.315156] env[63531]: DEBUG oslo_vmware.api [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118144, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.461699} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.315420] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 06ec91b2-14cb-4db8-8fab-e6c4183b1116/06ec91b2-14cb-4db8-8fab-e6c4183b1116.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 914.315667] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 06ec91b2-14cb-4db8-8fab-e6c4183b1116] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 914.315924] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b03d8fbd-b971-4c95-a05d-45e22c17a6f1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.317864] env[63531]: DEBUG oslo_concurrency.lockutils [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Releasing lock "refresh_cache-4b5e5821-9def-4b53-be19-355d9e5f81ec" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.318147] env[63531]: DEBUG nova.compute.manager [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Instance network_info: |[{"id": "d31275a8-8fba-48c3-be21-3b90d7cc05cf", "address": "fa:16:3e:b7:c3:f7", "network": {"id": "887fff70-ea19-458d-a945-e4f3ed6dc185", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2099566288-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "04a4c814245346a9b73253a25a822fef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49a4d142-3f97-47fe-b074-58923c46815e", "external-id": "nsx-vlan-transportzone-565", "segmentation_id": 565, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd31275a8-8f", "ovs_interfaceid": "d31275a8-8fba-48c3-be21-3b90d7cc05cf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 914.318415] env[63531]: DEBUG oslo_concurrency.lockutils [req-305bab81-7eb4-447d-ad2a-8e627c10b93c req-824cbf59-787d-417d-8208-5c9086fdba37 service nova] Acquired lock "refresh_cache-4b5e5821-9def-4b53-be19-355d9e5f81ec" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.318593] env[63531]: DEBUG nova.network.neutron [req-305bab81-7eb4-447d-ad2a-8e627c10b93c req-824cbf59-787d-417d-8208-5c9086fdba37 service nova] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Refreshing network info cache for port d31275a8-8fba-48c3-be21-3b90d7cc05cf {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 914.319742] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b7:c3:f7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '49a4d142-3f97-47fe-b074-58923c46815e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd31275a8-8fba-48c3-be21-3b90d7cc05cf', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 914.327244] env[63531]: DEBUG oslo.service.loopingcall [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 914.328537] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 914.328753] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1da7ad66-062f-4551-9cec-7c6f5f86fab0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.345677] env[63531]: DEBUG oslo_vmware.api [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 914.345677] env[63531]: value = "task-1118146" [ 914.345677] env[63531]: _type = "Task" [ 914.345677] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.351906] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 914.351906] env[63531]: value = "task-1118147" [ 914.351906] env[63531]: _type = "Task" [ 914.351906] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.355441] env[63531]: DEBUG oslo_vmware.api [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118146, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.363509] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118147, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.453020] env[63531]: DEBUG nova.scheduler.client.report [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 914.531012] env[63531]: DEBUG oslo_vmware.api [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118145, 'name': Rename_Task, 'duration_secs': 0.172824} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.531012] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 914.531012] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b39b0b2e-3d8b-4fc3-ae63-923e13f34848 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.539858] env[63531]: DEBUG oslo_vmware.api [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 914.539858] env[63531]: value = "task-1118148" [ 914.539858] env[63531]: _type = "Task" [ 914.539858] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.548892] env[63531]: DEBUG oslo_vmware.api [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118148, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.857016] env[63531]: DEBUG oslo_vmware.api [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118146, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069903} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.860174] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 06ec91b2-14cb-4db8-8fab-e6c4183b1116] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 914.863071] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d62ca888-ff92-484a-8e0c-25a963850b2c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.870716] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118147, 'name': CreateVM_Task, 'duration_secs': 0.389458} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.879520] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 914.888038] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 06ec91b2-14cb-4db8-8fab-e6c4183b1116] Reconfiguring VM instance instance-0000004e to attach disk [datastore1] 06ec91b2-14cb-4db8-8fab-e6c4183b1116/06ec91b2-14cb-4db8-8fab-e6c4183b1116.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 914.888732] env[63531]: DEBUG oslo_concurrency.lockutils [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.888904] env[63531]: DEBUG oslo_concurrency.lockutils [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.889234] env[63531]: DEBUG oslo_concurrency.lockutils [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 914.889456] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-199d2169-ec92-40aa-9243-1be550b0faad {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.905141] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6436fcb0-6bdb-40d4-b132-99a8744a1d25 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.910285] env[63531]: DEBUG oslo_vmware.api [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Waiting for the task: (returnval){ [ 914.910285] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]525bed46-301d-06a0-fac4-f12f11d40655" [ 914.910285] env[63531]: _type = "Task" [ 914.910285] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.915957] env[63531]: DEBUG oslo_vmware.api [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 914.915957] env[63531]: value = "task-1118149" [ 914.915957] env[63531]: _type = "Task" [ 914.915957] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.937864] env[63531]: DEBUG oslo_vmware.api [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]525bed46-301d-06a0-fac4-f12f11d40655, 'name': SearchDatastore_Task, 'duration_secs': 0.021614} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.938187] env[63531]: DEBUG oslo_concurrency.lockutils [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.938432] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 914.938707] env[63531]: DEBUG oslo_concurrency.lockutils [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.938865] env[63531]: DEBUG oslo_concurrency.lockutils [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.939064] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 914.939345] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8a5317a8-80bb-4523-9eb9-85e7141b74f8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.951174] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 914.951388] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 914.952276] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5e11ae30-0a0c-4ab9-8d36-2eb47afcb165 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.955249] env[63531]: DEBUG oslo_concurrency.lockutils [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.382s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.955388] env[63531]: INFO nova.compute.manager [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Migrating [ 914.963283] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.132s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.963636] env[63531]: DEBUG nova.objects.instance [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Lazy-loading 'pci_requests' on Instance uuid 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 914.974988] env[63531]: DEBUG oslo_vmware.api [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Waiting for the task: (returnval){ [ 914.974988] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52cf6f49-9fd7-e1dd-865d-c0f5fa5fd36e" [ 914.974988] env[63531]: _type = "Task" [ 914.974988] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.991055] env[63531]: DEBUG oslo_vmware.api [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52cf6f49-9fd7-e1dd-865d-c0f5fa5fd36e, 'name': SearchDatastore_Task, 'duration_secs': 0.010179} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.991824] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b5862d63-6665-4246-8a5d-c216bb6f64f8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.997946] env[63531]: DEBUG oslo_vmware.api [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Waiting for the task: (returnval){ [ 914.997946] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52e0ae90-9c4e-397c-4172-4b1797e05fa7" [ 914.997946] env[63531]: _type = "Task" [ 914.997946] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.006234] env[63531]: DEBUG oslo_vmware.api [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52e0ae90-9c4e-397c-4172-4b1797e05fa7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.051351] env[63531]: DEBUG oslo_vmware.api [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118148, 'name': PowerOnVM_Task} progress is 87%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.058124] env[63531]: DEBUG nova.network.neutron [req-305bab81-7eb4-447d-ad2a-8e627c10b93c req-824cbf59-787d-417d-8208-5c9086fdba37 service nova] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Updated VIF entry in instance network info cache for port d31275a8-8fba-48c3-be21-3b90d7cc05cf. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 915.058124] env[63531]: DEBUG nova.network.neutron [req-305bab81-7eb4-447d-ad2a-8e627c10b93c req-824cbf59-787d-417d-8208-5c9086fdba37 service nova] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Updating instance_info_cache with network_info: [{"id": "d31275a8-8fba-48c3-be21-3b90d7cc05cf", "address": "fa:16:3e:b7:c3:f7", "network": {"id": "887fff70-ea19-458d-a945-e4f3ed6dc185", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2099566288-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "04a4c814245346a9b73253a25a822fef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49a4d142-3f97-47fe-b074-58923c46815e", "external-id": "nsx-vlan-transportzone-565", "segmentation_id": 565, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd31275a8-8f", "ovs_interfaceid": "d31275a8-8fba-48c3-be21-3b90d7cc05cf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.426880] env[63531]: DEBUG oslo_vmware.api [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118149, 'name': ReconfigVM_Task, 'duration_secs': 0.305325} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.427244] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 06ec91b2-14cb-4db8-8fab-e6c4183b1116] Reconfigured VM instance instance-0000004e to attach disk [datastore1] 06ec91b2-14cb-4db8-8fab-e6c4183b1116/06ec91b2-14cb-4db8-8fab-e6c4183b1116.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 915.427870] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c3274fbb-7eca-42da-98fd-f5d8b20518e7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.437113] env[63531]: DEBUG oslo_vmware.api [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 915.437113] env[63531]: value = "task-1118150" [ 915.437113] env[63531]: _type = "Task" [ 915.437113] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.448624] env[63531]: DEBUG oslo_vmware.api [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118150, 'name': Rename_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.476818] env[63531]: DEBUG nova.objects.instance [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Lazy-loading 'numa_topology' on Instance uuid 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 915.478381] env[63531]: DEBUG oslo_concurrency.lockutils [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquiring lock "refresh_cache-9a3fad46-a15a-451c-bdab-a3c8cc8add07" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.478573] env[63531]: DEBUG oslo_concurrency.lockutils [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquired lock "refresh_cache-9a3fad46-a15a-451c-bdab-a3c8cc8add07" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.478768] env[63531]: DEBUG nova.network.neutron [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 915.509103] env[63531]: DEBUG oslo_vmware.api [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52e0ae90-9c4e-397c-4172-4b1797e05fa7, 'name': SearchDatastore_Task, 'duration_secs': 0.009878} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.509345] env[63531]: DEBUG oslo_concurrency.lockutils [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.509610] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 4b5e5821-9def-4b53-be19-355d9e5f81ec/4b5e5821-9def-4b53-be19-355d9e5f81ec.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 915.509957] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-21cdb0a3-ec62-499a-9782-f28c9d7b8d7c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.517751] env[63531]: DEBUG oslo_vmware.api [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Waiting for the task: (returnval){ [ 915.517751] env[63531]: value = "task-1118151" [ 915.517751] env[63531]: _type = "Task" [ 915.517751] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.527322] env[63531]: DEBUG oslo_vmware.api [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1118151, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.551687] env[63531]: DEBUG oslo_vmware.api [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118148, 'name': PowerOnVM_Task, 'duration_secs': 0.908537} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.552111] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 915.552475] env[63531]: INFO nova.compute.manager [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Took 8.98 seconds to spawn the instance on the hypervisor. [ 915.552713] env[63531]: DEBUG nova.compute.manager [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 915.553626] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-867dac9d-bcd9-4891-8c38-c0ae45c91112 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.564381] env[63531]: DEBUG oslo_concurrency.lockutils [req-305bab81-7eb4-447d-ad2a-8e627c10b93c req-824cbf59-787d-417d-8208-5c9086fdba37 service nova] Releasing lock "refresh_cache-4b5e5821-9def-4b53-be19-355d9e5f81ec" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.949516] env[63531]: DEBUG oslo_vmware.api [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118150, 'name': Rename_Task, 'duration_secs': 0.163347} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.949691] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 06ec91b2-14cb-4db8-8fab-e6c4183b1116] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 915.953019] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c0d1d9b3-5b37-44ce-b869-2d0f45860a9c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.963472] env[63531]: DEBUG oslo_vmware.api [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 915.963472] env[63531]: value = "task-1118152" [ 915.963472] env[63531]: _type = "Task" [ 915.963472] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.978123] env[63531]: DEBUG oslo_vmware.api [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118152, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.981423] env[63531]: INFO nova.compute.claims [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 916.033370] env[63531]: DEBUG oslo_vmware.api [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1118151, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.075092] env[63531]: INFO nova.compute.manager [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Took 27.56 seconds to build instance. [ 916.234065] env[63531]: DEBUG nova.network.neutron [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Updating instance_info_cache with network_info: [{"id": "e4133e0f-55c1-4b3f-847d-899dd1167eec", "address": "fa:16:3e:90:40:9b", "network": {"id": "c4f4c994-36ab-42ab-94f8-78f29cbafa50", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-994507698-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "57e9bf4ee0f1463fa67b0f8deff30fdc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "767a3a48-41d4-4a0c-961d-0024837f63bd", "external-id": "nsx-vlan-transportzone-141", "segmentation_id": 141, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4133e0f-55", "ovs_interfaceid": "e4133e0f-55c1-4b3f-847d-899dd1167eec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.474014] env[63531]: DEBUG oslo_vmware.api [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118152, 'name': PowerOnVM_Task, 'duration_secs': 0.491842} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.474389] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 06ec91b2-14cb-4db8-8fab-e6c4183b1116] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 916.474509] env[63531]: INFO nova.compute.manager [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 06ec91b2-14cb-4db8-8fab-e6c4183b1116] Took 7.41 seconds to spawn the instance on the hypervisor. [ 916.474699] env[63531]: DEBUG nova.compute.manager [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 06ec91b2-14cb-4db8-8fab-e6c4183b1116] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 916.475494] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c8a2052-3972-4eb9-b56c-3733751aa814 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.530324] env[63531]: DEBUG oslo_vmware.api [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1118151, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.528931} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.530608] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 4b5e5821-9def-4b53-be19-355d9e5f81ec/4b5e5821-9def-4b53-be19-355d9e5f81ec.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 916.530828] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 916.531099] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-52e7934b-7459-4e5e-ac23-fe9d280fb9b6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.539506] env[63531]: DEBUG oslo_vmware.api [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Waiting for the task: (returnval){ [ 916.539506] env[63531]: value = "task-1118153" [ 916.539506] env[63531]: _type = "Task" [ 916.539506] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.548013] env[63531]: DEBUG oslo_vmware.api [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1118153, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.577211] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5ce096d4-3e9d-417e-9587-69e08c5fcee4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "6118e242-c0db-4707-a5f9-3d12b823935f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.077s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.737299] env[63531]: DEBUG oslo_concurrency.lockutils [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Releasing lock "refresh_cache-9a3fad46-a15a-451c-bdab-a3c8cc8add07" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.996728] env[63531]: INFO nova.compute.manager [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 06ec91b2-14cb-4db8-8fab-e6c4183b1116] Took 25.90 seconds to build instance. [ 917.051478] env[63531]: DEBUG oslo_vmware.api [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1118153, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066922} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.054177] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 917.055509] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02b1b351-a898-4ffe-be27-2dfb4ae1f47f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.079034] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Reconfiguring VM instance instance-0000004f to attach disk [datastore1] 4b5e5821-9def-4b53-be19-355d9e5f81ec/4b5e5821-9def-4b53-be19-355d9e5f81ec.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 917.084981] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e206a805-83dd-406d-975a-9468ee2c7bc0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.105990] env[63531]: DEBUG oslo_vmware.api [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Waiting for the task: (returnval){ [ 917.105990] env[63531]: value = "task-1118154" [ 917.105990] env[63531]: _type = "Task" [ 917.105990] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.117019] env[63531]: DEBUG oslo_vmware.api [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1118154, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.334017] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2a299d2-462d-4ba2-bdd9-9e8aa032827d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.340128] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1daa941d-b152-4eaf-88bf-a389265196ee {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.371164] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da406c22-f4eb-4a05-938b-622ac4974858 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.378398] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92a1a847-9ec0-41a1-986d-900eccfe7a27 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.391913] env[63531]: DEBUG nova.compute.provider_tree [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 917.499527] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d080b7f-b97f-4300-ad68-815c9883d6f4 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "06ec91b2-14cb-4db8-8fab-e6c4183b1116" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.423s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.615825] env[63531]: DEBUG oslo_vmware.api [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1118154, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.657206] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4fccfddd-dc56-46bb-aba7-1e6242d29900 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "06ec91b2-14cb-4db8-8fab-e6c4183b1116" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.657206] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4fccfddd-dc56-46bb-aba7-1e6242d29900 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "06ec91b2-14cb-4db8-8fab-e6c4183b1116" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.657206] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4fccfddd-dc56-46bb-aba7-1e6242d29900 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "06ec91b2-14cb-4db8-8fab-e6c4183b1116-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.657206] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4fccfddd-dc56-46bb-aba7-1e6242d29900 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "06ec91b2-14cb-4db8-8fab-e6c4183b1116-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.657465] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4fccfddd-dc56-46bb-aba7-1e6242d29900 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "06ec91b2-14cb-4db8-8fab-e6c4183b1116-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.659535] env[63531]: INFO nova.compute.manager [None req-4fccfddd-dc56-46bb-aba7-1e6242d29900 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 06ec91b2-14cb-4db8-8fab-e6c4183b1116] Terminating instance [ 917.661285] env[63531]: DEBUG nova.compute.manager [None req-4fccfddd-dc56-46bb-aba7-1e6242d29900 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 06ec91b2-14cb-4db8-8fab-e6c4183b1116] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 917.661480] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-4fccfddd-dc56-46bb-aba7-1e6242d29900 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 06ec91b2-14cb-4db8-8fab-e6c4183b1116] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 917.662262] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-802a1b34-5e07-4e0e-9458-17973bb9b105 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.669501] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-4fccfddd-dc56-46bb-aba7-1e6242d29900 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 06ec91b2-14cb-4db8-8fab-e6c4183b1116] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 917.669717] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5ea4caff-3064-45ed-ac80-9e11a2caa7cd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.675715] env[63531]: DEBUG oslo_vmware.api [None req-4fccfddd-dc56-46bb-aba7-1e6242d29900 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 917.675715] env[63531]: value = "task-1118155" [ 917.675715] env[63531]: _type = "Task" [ 917.675715] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.682557] env[63531]: DEBUG oslo_vmware.api [None req-4fccfddd-dc56-46bb-aba7-1e6242d29900 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118155, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.727325] env[63531]: INFO nova.compute.manager [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Rebuilding instance [ 917.764912] env[63531]: DEBUG nova.compute.manager [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 917.765915] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f41b492-b01a-497e-9489-c609eaed1680 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.895779] env[63531]: DEBUG nova.scheduler.client.report [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 918.115721] env[63531]: DEBUG oslo_vmware.api [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1118154, 'name': ReconfigVM_Task, 'duration_secs': 0.706206} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.116021] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Reconfigured VM instance instance-0000004f to attach disk [datastore1] 4b5e5821-9def-4b53-be19-355d9e5f81ec/4b5e5821-9def-4b53-be19-355d9e5f81ec.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 918.116797] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-aace5726-3cf3-49f8-a39e-26ad930dd42e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.123422] env[63531]: DEBUG oslo_vmware.api [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Waiting for the task: (returnval){ [ 918.123422] env[63531]: value = "task-1118156" [ 918.123422] env[63531]: _type = "Task" [ 918.123422] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.131020] env[63531]: DEBUG oslo_vmware.api [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1118156, 'name': Rename_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.185377] env[63531]: DEBUG oslo_vmware.api [None req-4fccfddd-dc56-46bb-aba7-1e6242d29900 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118155, 'name': PowerOffVM_Task, 'duration_secs': 0.231003} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.185726] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-4fccfddd-dc56-46bb-aba7-1e6242d29900 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 06ec91b2-14cb-4db8-8fab-e6c4183b1116] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 918.185908] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-4fccfddd-dc56-46bb-aba7-1e6242d29900 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 06ec91b2-14cb-4db8-8fab-e6c4183b1116] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 918.186179] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a5a68627-7089-45c8-9669-6cdecbcb43a7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.251281] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bf7b753-f0d4-4a5e-87df-096ac5a9856a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.271212] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Updating instance '9a3fad46-a15a-451c-bdab-a3c8cc8add07' progress to 0 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 918.278702] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 918.279071] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-72320590-7d80-4e5c-ae2d-30a6ca7d684e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.286253] env[63531]: DEBUG oslo_vmware.api [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 918.286253] env[63531]: value = "task-1118158" [ 918.286253] env[63531]: _type = "Task" [ 918.286253] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.296663] env[63531]: DEBUG oslo_vmware.api [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118158, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.402154] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.439s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.404453] env[63531]: DEBUG oslo_concurrency.lockutils [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.286s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.406528] env[63531]: INFO nova.compute.claims [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 918.437545] env[63531]: INFO nova.network.neutron [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Updating port 2d27a8a5-c288-44fe-ac09-0d3b488386d6 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 918.528022] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-4fccfddd-dc56-46bb-aba7-1e6242d29900 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 06ec91b2-14cb-4db8-8fab-e6c4183b1116] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 918.528532] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-4fccfddd-dc56-46bb-aba7-1e6242d29900 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 06ec91b2-14cb-4db8-8fab-e6c4183b1116] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 918.528532] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-4fccfddd-dc56-46bb-aba7-1e6242d29900 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Deleting the datastore file [datastore1] 06ec91b2-14cb-4db8-8fab-e6c4183b1116 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 918.528733] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a86dd4ce-51a3-404f-8a78-af36ff84f03e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.536265] env[63531]: DEBUG oslo_vmware.api [None req-4fccfddd-dc56-46bb-aba7-1e6242d29900 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 918.536265] env[63531]: value = "task-1118159" [ 918.536265] env[63531]: _type = "Task" [ 918.536265] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.544580] env[63531]: DEBUG oslo_vmware.api [None req-4fccfddd-dc56-46bb-aba7-1e6242d29900 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118159, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.635443] env[63531]: DEBUG oslo_vmware.api [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1118156, 'name': Rename_Task, 'duration_secs': 0.136918} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.635845] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 918.636159] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9a7fe20c-2bdf-43c4-9114-0791616567e4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.642879] env[63531]: DEBUG oslo_vmware.api [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Waiting for the task: (returnval){ [ 918.642879] env[63531]: value = "task-1118160" [ 918.642879] env[63531]: _type = "Task" [ 918.642879] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.651469] env[63531]: DEBUG oslo_vmware.api [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1118160, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.777534] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 918.777946] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-407e8219-6b8c-4bf4-ba65-871fc332a881 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.785808] env[63531]: DEBUG oslo_vmware.api [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 918.785808] env[63531]: value = "task-1118161" [ 918.785808] env[63531]: _type = "Task" [ 918.785808] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.802196] env[63531]: DEBUG oslo_vmware.api [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118158, 'name': PowerOffVM_Task, 'duration_secs': 0.256943} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.802673] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 918.802993] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 918.803367] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] VM already powered off {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 918.803652] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Updating instance '9a3fad46-a15a-451c-bdab-a3c8cc8add07' progress to 17 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 918.808971] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18d40098-e4f5-4000-9d50-006dfd524f43 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.818461] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 918.818785] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dd36162e-4777-4d17-a5b5-aa34ff100eca {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.886218] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 918.886525] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 918.886788] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Deleting the datastore file [datastore1] 6118e242-c0db-4707-a5f9-3d12b823935f {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 918.887410] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5123aa72-402a-4638-83be-6650e472df12 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.894066] env[63531]: DEBUG oslo_vmware.api [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 918.894066] env[63531]: value = "task-1118163" [ 918.894066] env[63531]: _type = "Task" [ 918.894066] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.902241] env[63531]: DEBUG oslo_vmware.api [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118163, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.047026] env[63531]: DEBUG oslo_vmware.api [None req-4fccfddd-dc56-46bb-aba7-1e6242d29900 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118159, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.122523} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.047352] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-4fccfddd-dc56-46bb-aba7-1e6242d29900 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 919.048528] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-4fccfddd-dc56-46bb-aba7-1e6242d29900 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 06ec91b2-14cb-4db8-8fab-e6c4183b1116] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 919.048528] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-4fccfddd-dc56-46bb-aba7-1e6242d29900 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 06ec91b2-14cb-4db8-8fab-e6c4183b1116] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 919.048528] env[63531]: INFO nova.compute.manager [None req-4fccfddd-dc56-46bb-aba7-1e6242d29900 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 06ec91b2-14cb-4db8-8fab-e6c4183b1116] Took 1.39 seconds to destroy the instance on the hypervisor. [ 919.048528] env[63531]: DEBUG oslo.service.loopingcall [None req-4fccfddd-dc56-46bb-aba7-1e6242d29900 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 919.048528] env[63531]: DEBUG nova.compute.manager [-] [instance: 06ec91b2-14cb-4db8-8fab-e6c4183b1116] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 919.048528] env[63531]: DEBUG nova.network.neutron [-] [instance: 06ec91b2-14cb-4db8-8fab-e6c4183b1116] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 919.153446] env[63531]: DEBUG oslo_vmware.api [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1118160, 'name': PowerOnVM_Task, 'duration_secs': 0.500607} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.153729] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 919.153935] env[63531]: INFO nova.compute.manager [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Took 7.63 seconds to spawn the instance on the hypervisor. [ 919.154133] env[63531]: DEBUG nova.compute.manager [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 919.154876] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7aad36d-76b4-4b08-a493-71d7aab87d97 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.306254] env[63531]: DEBUG nova.compute.manager [req-e4975167-56d5-4f8d-b512-e36df4cf974b req-e0954f0d-0d43-4bb4-84d4-7709773a06a8 service nova] [instance: 06ec91b2-14cb-4db8-8fab-e6c4183b1116] Received event network-vif-deleted-5c98d4e0-4d30-4104-8ee9-15361087529a {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 919.306363] env[63531]: INFO nova.compute.manager [req-e4975167-56d5-4f8d-b512-e36df4cf974b req-e0954f0d-0d43-4bb4-84d4-7709773a06a8 service nova] [instance: 06ec91b2-14cb-4db8-8fab-e6c4183b1116] Neutron deleted interface 5c98d4e0-4d30-4104-8ee9-15361087529a; detaching it from the instance and deleting it from the info cache [ 919.306549] env[63531]: DEBUG nova.network.neutron [req-e4975167-56d5-4f8d-b512-e36df4cf974b req-e0954f0d-0d43-4bb4-84d4-7709773a06a8 service nova] [instance: 06ec91b2-14cb-4db8-8fab-e6c4183b1116] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.316478] env[63531]: DEBUG nova.virt.hardware [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 919.316707] env[63531]: DEBUG nova.virt.hardware [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 919.316872] env[63531]: DEBUG nova.virt.hardware [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 919.317077] env[63531]: DEBUG nova.virt.hardware [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 919.317237] env[63531]: DEBUG nova.virt.hardware [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 919.317390] env[63531]: DEBUG nova.virt.hardware [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 919.317630] env[63531]: DEBUG nova.virt.hardware [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 919.317803] env[63531]: DEBUG nova.virt.hardware [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 919.317978] env[63531]: DEBUG nova.virt.hardware [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 919.318164] env[63531]: DEBUG nova.virt.hardware [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 919.318343] env[63531]: DEBUG nova.virt.hardware [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 919.323838] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d33dd8dc-d37b-4fb2-84bb-e683b3d67828 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.339786] env[63531]: DEBUG oslo_vmware.api [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 919.339786] env[63531]: value = "task-1118164" [ 919.339786] env[63531]: _type = "Task" [ 919.339786] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.347431] env[63531]: DEBUG oslo_vmware.api [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118164, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.403490] env[63531]: DEBUG oslo_vmware.api [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118163, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.135088} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.403803] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 919.404334] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 919.404334] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 919.663842] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6645510a-6eb6-40a0-bb5f-7c9990745dff {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.674949] env[63531]: INFO nova.compute.manager [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Took 27.48 seconds to build instance. [ 919.676740] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d27701e-4d24-4a58-9046-5feffe08f1c5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.709620] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1afe3710-85b4-49bf-845a-e8262f2a660c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.717814] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8df61f6-b0e7-4316-bb2a-034e70dedf6f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.731521] env[63531]: DEBUG nova.compute.provider_tree [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 919.785659] env[63531]: DEBUG nova.network.neutron [-] [instance: 06ec91b2-14cb-4db8-8fab-e6c4183b1116] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.809378] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1eb91a11-9b41-4fae-8165-d1cef2a33f5a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.818297] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57cf8232-1e86-44c8-8c6a-ef71da6650c1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.848261] env[63531]: DEBUG nova.compute.manager [req-e4975167-56d5-4f8d-b512-e36df4cf974b req-e0954f0d-0d43-4bb4-84d4-7709773a06a8 service nova] [instance: 06ec91b2-14cb-4db8-8fab-e6c4183b1116] Detach interface failed, port_id=5c98d4e0-4d30-4104-8ee9-15361087529a, reason: Instance 06ec91b2-14cb-4db8-8fab-e6c4183b1116 could not be found. {{(pid=63531) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 919.856232] env[63531]: DEBUG oslo_vmware.api [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118164, 'name': ReconfigVM_Task, 'duration_secs': 0.132694} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.857182] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Updating instance '9a3fad46-a15a-451c-bdab-a3c8cc8add07' progress to 33 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 920.184085] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Acquiring lock "refresh_cache-2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.184085] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Acquired lock "refresh_cache-2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.184085] env[63531]: DEBUG nova.network.neutron [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 920.184856] env[63531]: DEBUG oslo_concurrency.lockutils [None req-75ea813b-f748-4b22-bb4f-9d8c1faffc4f tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Lock "4b5e5821-9def-4b53-be19-355d9e5f81ec" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.998s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.234159] env[63531]: DEBUG nova.scheduler.client.report [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 920.288472] env[63531]: INFO nova.compute.manager [-] [instance: 06ec91b2-14cb-4db8-8fab-e6c4183b1116] Took 1.24 seconds to deallocate network for instance. [ 920.362221] env[63531]: DEBUG nova.virt.hardware [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 920.362514] env[63531]: DEBUG nova.virt.hardware [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 920.362723] env[63531]: DEBUG nova.virt.hardware [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 920.362931] env[63531]: DEBUG nova.virt.hardware [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 920.363103] env[63531]: DEBUG nova.virt.hardware [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 920.363258] env[63531]: DEBUG nova.virt.hardware [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 920.363467] env[63531]: DEBUG nova.virt.hardware [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 920.363628] env[63531]: DEBUG nova.virt.hardware [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 920.363799] env[63531]: DEBUG nova.virt.hardware [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 920.363999] env[63531]: DEBUG nova.virt.hardware [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 920.364169] env[63531]: DEBUG nova.virt.hardware [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 920.369969] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Reconfiguring VM instance instance-00000036 to detach disk 2000 {{(pid=63531) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 920.370268] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4ee028ec-1763-4d06-b658-83aa3675400b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.388836] env[63531]: DEBUG oslo_vmware.api [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 920.388836] env[63531]: value = "task-1118165" [ 920.388836] env[63531]: _type = "Task" [ 920.388836] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.396896] env[63531]: DEBUG oslo_vmware.api [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118165, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.440699] env[63531]: DEBUG nova.virt.hardware [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 920.441050] env[63531]: DEBUG nova.virt.hardware [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 920.441246] env[63531]: DEBUG nova.virt.hardware [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 920.441441] env[63531]: DEBUG nova.virt.hardware [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 920.441598] env[63531]: DEBUG nova.virt.hardware [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 920.441754] env[63531]: DEBUG nova.virt.hardware [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 920.441972] env[63531]: DEBUG nova.virt.hardware [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 920.442153] env[63531]: DEBUG nova.virt.hardware [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 920.442328] env[63531]: DEBUG nova.virt.hardware [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 920.442493] env[63531]: DEBUG nova.virt.hardware [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 920.442672] env[63531]: DEBUG nova.virt.hardware [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 920.443637] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cc979a7-f678-4116-8c4d-8f08fa1c1474 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.451497] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6225cab0-2979-43db-bb8e-ca70ffa9f437 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.465959] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:43:5d:33', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c8f580e6-1d86-41ee-9ebe-c531cb9299c6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0997f575-14bc-4813-9df6-b18e4c1c2186', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 920.473939] env[63531]: DEBUG oslo.service.loopingcall [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 920.474442] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 920.474561] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0e4488d2-9c65-4172-9818-0abcf308e2cd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.494788] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 920.494788] env[63531]: value = "task-1118166" [ 920.494788] env[63531]: _type = "Task" [ 920.494788] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.502755] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118166, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.739266] env[63531]: DEBUG oslo_concurrency.lockutils [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.335s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.739793] env[63531]: DEBUG nova.compute.manager [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 920.742461] env[63531]: DEBUG oslo_concurrency.lockutils [None req-935264d7-e19f-4031-a0a0-6f80fb185255 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.362s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.742631] env[63531]: DEBUG nova.objects.instance [None req-935264d7-e19f-4031-a0a0-6f80fb185255 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lazy-loading 'resources' on Instance uuid e95d3273-8216-47cc-95b6-99301366a827 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 920.797374] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4fccfddd-dc56-46bb-aba7-1e6242d29900 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.901610] env[63531]: DEBUG oslo_vmware.api [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118165, 'name': ReconfigVM_Task, 'duration_secs': 0.151784} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.901773] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Reconfigured VM instance instance-00000036 to detach disk 2000 {{(pid=63531) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 920.902586] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3bb2600-471e-40f5-9b57-ee4f5c7ec5f1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.929651] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Reconfiguring VM instance instance-00000036 to attach disk [datastore1] 9a3fad46-a15a-451c-bdab-a3c8cc8add07/9a3fad46-a15a-451c-bdab-a3c8cc8add07.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 920.929976] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5f7b38f8-bcd1-4165-947f-8e12fa36d9dd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.946038] env[63531]: DEBUG nova.network.neutron [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Updating instance_info_cache with network_info: [{"id": "2d27a8a5-c288-44fe-ac09-0d3b488386d6", "address": "fa:16:3e:91:1c:62", "network": {"id": "5b8b7134-f123-40b8-9134-5cd072e8ad01", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1232565840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.165", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef3b56ce3f3a4155bd09642f5c48ea6f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2d27a8a5-c2", "ovs_interfaceid": "2d27a8a5-c288-44fe-ac09-0d3b488386d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 920.953129] env[63531]: DEBUG oslo_vmware.api [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 920.953129] env[63531]: value = "task-1118167" [ 920.953129] env[63531]: _type = "Task" [ 920.953129] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.963394] env[63531]: DEBUG oslo_vmware.api [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118167, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.004398] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118166, 'name': CreateVM_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.246744] env[63531]: DEBUG nova.compute.utils [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 921.250547] env[63531]: DEBUG nova.compute.manager [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Not allocating networking since 'none' was specified. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 921.361568] env[63531]: DEBUG nova.compute.manager [req-7e9e3cc9-43ba-4758-bf87-b5c3a6243411 req-ed25c8d5-44fd-4f04-84d4-f00ba98c41b9 service nova] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Received event network-vif-plugged-2d27a8a5-c288-44fe-ac09-0d3b488386d6 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 921.361681] env[63531]: DEBUG oslo_concurrency.lockutils [req-7e9e3cc9-43ba-4758-bf87-b5c3a6243411 req-ed25c8d5-44fd-4f04-84d4-f00ba98c41b9 service nova] Acquiring lock "2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.361882] env[63531]: DEBUG oslo_concurrency.lockutils [req-7e9e3cc9-43ba-4758-bf87-b5c3a6243411 req-ed25c8d5-44fd-4f04-84d4-f00ba98c41b9 service nova] Lock "2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.362061] env[63531]: DEBUG oslo_concurrency.lockutils [req-7e9e3cc9-43ba-4758-bf87-b5c3a6243411 req-ed25c8d5-44fd-4f04-84d4-f00ba98c41b9 service nova] Lock "2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.362245] env[63531]: DEBUG nova.compute.manager [req-7e9e3cc9-43ba-4758-bf87-b5c3a6243411 req-ed25c8d5-44fd-4f04-84d4-f00ba98c41b9 service nova] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] No waiting events found dispatching network-vif-plugged-2d27a8a5-c288-44fe-ac09-0d3b488386d6 {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 921.362416] env[63531]: WARNING nova.compute.manager [req-7e9e3cc9-43ba-4758-bf87-b5c3a6243411 req-ed25c8d5-44fd-4f04-84d4-f00ba98c41b9 service nova] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Received unexpected event network-vif-plugged-2d27a8a5-c288-44fe-ac09-0d3b488386d6 for instance with vm_state shelved_offloaded and task_state spawning. [ 921.362580] env[63531]: DEBUG nova.compute.manager [req-7e9e3cc9-43ba-4758-bf87-b5c3a6243411 req-ed25c8d5-44fd-4f04-84d4-f00ba98c41b9 service nova] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Received event network-changed-2d27a8a5-c288-44fe-ac09-0d3b488386d6 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 921.362737] env[63531]: DEBUG nova.compute.manager [req-7e9e3cc9-43ba-4758-bf87-b5c3a6243411 req-ed25c8d5-44fd-4f04-84d4-f00ba98c41b9 service nova] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Refreshing instance network info cache due to event network-changed-2d27a8a5-c288-44fe-ac09-0d3b488386d6. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 921.362907] env[63531]: DEBUG oslo_concurrency.lockutils [req-7e9e3cc9-43ba-4758-bf87-b5c3a6243411 req-ed25c8d5-44fd-4f04-84d4-f00ba98c41b9 service nova] Acquiring lock "refresh_cache-2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.448879] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Releasing lock "refresh_cache-2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.450359] env[63531]: DEBUG oslo_concurrency.lockutils [req-7e9e3cc9-43ba-4758-bf87-b5c3a6243411 req-ed25c8d5-44fd-4f04-84d4-f00ba98c41b9 service nova] Acquired lock "refresh_cache-2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.451066] env[63531]: DEBUG nova.network.neutron [req-7e9e3cc9-43ba-4758-bf87-b5c3a6243411 req-ed25c8d5-44fd-4f04-84d4-f00ba98c41b9 service nova] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Refreshing network info cache for port 2d27a8a5-c288-44fe-ac09-0d3b488386d6 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 921.463398] env[63531]: DEBUG oslo_vmware.api [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118167, 'name': ReconfigVM_Task, 'duration_secs': 0.251395} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.463703] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Reconfigured VM instance instance-00000036 to attach disk [datastore1] 9a3fad46-a15a-451c-bdab-a3c8cc8add07/9a3fad46-a15a-451c-bdab-a3c8cc8add07.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 921.463939] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Updating instance '9a3fad46-a15a-451c-bdab-a3c8cc8add07' progress to 50 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 921.478218] env[63531]: DEBUG nova.virt.hardware [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='f6da5bdb7ce96092ee5c84ce253ac9ac',container_format='bare',created_at=2024-10-03T08:21:29Z,direct_url=,disk_format='vmdk',id=d477826e-3e39-4ca7-bc24-998780166e1c,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-1173305590-shelved',owner='ef3b56ce3f3a4155bd09642f5c48ea6f',properties=ImageMetaProps,protected=,size=31666688,status='active',tags=,updated_at=2024-10-03T08:21:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 921.478477] env[63531]: DEBUG nova.virt.hardware [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 921.478667] env[63531]: DEBUG nova.virt.hardware [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 921.478882] env[63531]: DEBUG nova.virt.hardware [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 921.479075] env[63531]: DEBUG nova.virt.hardware [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 921.479273] env[63531]: DEBUG nova.virt.hardware [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 921.479525] env[63531]: DEBUG nova.virt.hardware [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 921.479710] env[63531]: DEBUG nova.virt.hardware [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 921.479884] env[63531]: DEBUG nova.virt.hardware [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 921.480064] env[63531]: DEBUG nova.virt.hardware [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 921.480285] env[63531]: DEBUG nova.virt.hardware [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 921.481128] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d914b9d-5eb8-45bb-b43d-9d7038b245f2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.491717] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e244b1f6-a3b4-4fee-9054-fad55121b565 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.509669] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:91:1c:62', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd891a662-7da0-4be5-8d0d-01b4ff70552a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2d27a8a5-c288-44fe-ac09-0d3b488386d6', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 921.516915] env[63531]: DEBUG oslo.service.loopingcall [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 921.519766] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 921.520171] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1eec71ca-1d21-4ee5-acce-aa5ab3233bed {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.538724] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118166, 'name': CreateVM_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.539390] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49967617-4112-43e8-8ef1-d9f8c27b9cd9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.545571] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-879e725f-e4b3-4416-999d-e18154f39794 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.549445] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 921.549445] env[63531]: value = "task-1118168" [ 921.549445] env[63531]: _type = "Task" [ 921.549445] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.576682] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b24d0ef9-896d-44f5-8c28-1a384cb28525 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.582030] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118168, 'name': CreateVM_Task} progress is 15%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.586570] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5cba4df-f3e3-44a4-a6b1-5ff5d2e92757 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.599853] env[63531]: DEBUG nova.compute.provider_tree [None req-935264d7-e19f-4031-a0a0-6f80fb185255 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 921.634631] env[63531]: DEBUG oslo_concurrency.lockutils [None req-10dd9342-244a-47e0-b955-67751e3d781a tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Acquiring lock "f46d8a85-6ab7-436a-bc5e-40d0f80974e5" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.634631] env[63531]: DEBUG oslo_concurrency.lockutils [None req-10dd9342-244a-47e0-b955-67751e3d781a tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lock "f46d8a85-6ab7-436a-bc5e-40d0f80974e5" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.751013] env[63531]: DEBUG nova.compute.manager [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 921.769595] env[63531]: DEBUG nova.compute.manager [None req-05439101-9bc6-4c60-84fa-463c67262c69 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 921.770681] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c9c2e04-516a-4f7c-b0bb-6c4ebe75fa1a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.972926] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b485d756-8796-41f7-b5f0-f7507f8e09ea {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.999326] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc275bb8-16b0-4731-ac99-65291bce7443 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.020912] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Updating instance '9a3fad46-a15a-451c-bdab-a3c8cc8add07' progress to 67 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 922.029024] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118166, 'name': CreateVM_Task, 'duration_secs': 1.181856} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.029024] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 922.029024] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.029024] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.029024] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 922.029308] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7988bd54-9097-451e-a483-0aeed77ba0dc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.034074] env[63531]: DEBUG oslo_vmware.api [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 922.034074] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5256011c-5694-9dcc-acc1-9c6feaeab8d6" [ 922.034074] env[63531]: _type = "Task" [ 922.034074] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.043020] env[63531]: DEBUG oslo_vmware.api [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5256011c-5694-9dcc-acc1-9c6feaeab8d6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.059181] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118168, 'name': CreateVM_Task, 'duration_secs': 0.333169} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.059369] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 922.061398] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d477826e-3e39-4ca7-bc24-998780166e1c" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.061398] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d477826e-3e39-4ca7-bc24-998780166e1c" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.061398] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d477826e-3e39-4ca7-bc24-998780166e1c" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 922.061398] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bc9595df-9a4a-4a76-ae5e-6fe81187f0ab {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.065880] env[63531]: DEBUG oslo_vmware.api [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Waiting for the task: (returnval){ [ 922.065880] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5252af36-edf6-8c38-b340-bb527fcf9219" [ 922.065880] env[63531]: _type = "Task" [ 922.065880] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.074994] env[63531]: DEBUG oslo_vmware.api [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5252af36-edf6-8c38-b340-bb527fcf9219, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.102478] env[63531]: DEBUG nova.scheduler.client.report [None req-935264d7-e19f-4031-a0a0-6f80fb185255 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 922.136921] env[63531]: INFO nova.compute.manager [None req-10dd9342-244a-47e0-b955-67751e3d781a tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Detaching volume c89ce39a-7d3d-468e-b6d6-7df150ed8817 [ 922.179205] env[63531]: INFO nova.virt.block_device [None req-10dd9342-244a-47e0-b955-67751e3d781a tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Attempting to driver detach volume c89ce39a-7d3d-468e-b6d6-7df150ed8817 from mountpoint /dev/sdb [ 922.179598] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-10dd9342-244a-47e0-b955-67751e3d781a tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Volume detach. Driver type: vmdk {{(pid=63531) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 922.179671] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-10dd9342-244a-47e0-b955-67751e3d781a tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244725', 'volume_id': 'c89ce39a-7d3d-468e-b6d6-7df150ed8817', 'name': 'volume-c89ce39a-7d3d-468e-b6d6-7df150ed8817', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f46d8a85-6ab7-436a-bc5e-40d0f80974e5', 'attached_at': '', 'detached_at': '', 'volume_id': 'c89ce39a-7d3d-468e-b6d6-7df150ed8817', 'serial': 'c89ce39a-7d3d-468e-b6d6-7df150ed8817'} {{(pid=63531) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 922.180605] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfb8780f-7299-4d35-a7a9-37ff9570159d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.207367] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68d39406-0321-44f0-a5e7-47ae5106761a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.215062] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5bf0081-a520-4a3b-b6c8-48c089068fa8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.236438] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5acd72c4-8e6c-47dd-ba32-01b59fc75759 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.252221] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4b9a0fd3-15d1-440d-90b4-9ce5235e55e0 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquiring lock "37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.252584] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4b9a0fd3-15d1-440d-90b4-9ce5235e55e0 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.252918] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4b9a0fd3-15d1-440d-90b4-9ce5235e55e0 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquiring lock "37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.253185] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4b9a0fd3-15d1-440d-90b4-9ce5235e55e0 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.253373] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4b9a0fd3-15d1-440d-90b4-9ce5235e55e0 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.255042] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-10dd9342-244a-47e0-b955-67751e3d781a tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] The volume has not been displaced from its original location: [datastore1] volume-c89ce39a-7d3d-468e-b6d6-7df150ed8817/volume-c89ce39a-7d3d-468e-b6d6-7df150ed8817.vmdk. No consolidation needed. {{(pid=63531) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 922.260526] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-10dd9342-244a-47e0-b955-67751e3d781a tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Reconfiguring VM instance instance-0000003c to detach disk 2001 {{(pid=63531) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 922.261135] env[63531]: INFO nova.compute.manager [None req-4b9a0fd3-15d1-440d-90b4-9ce5235e55e0 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Terminating instance [ 922.263165] env[63531]: DEBUG nova.network.neutron [req-7e9e3cc9-43ba-4758-bf87-b5c3a6243411 req-ed25c8d5-44fd-4f04-84d4-f00ba98c41b9 service nova] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Updated VIF entry in instance network info cache for port 2d27a8a5-c288-44fe-ac09-0d3b488386d6. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 922.263543] env[63531]: DEBUG nova.network.neutron [req-7e9e3cc9-43ba-4758-bf87-b5c3a6243411 req-ed25c8d5-44fd-4f04-84d4-f00ba98c41b9 service nova] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Updating instance_info_cache with network_info: [{"id": "2d27a8a5-c288-44fe-ac09-0d3b488386d6", "address": "fa:16:3e:91:1c:62", "network": {"id": "5b8b7134-f123-40b8-9134-5cd072e8ad01", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1232565840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.165", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef3b56ce3f3a4155bd09642f5c48ea6f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2d27a8a5-c2", "ovs_interfaceid": "2d27a8a5-c288-44fe-ac09-0d3b488386d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 922.267629] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a771d0d6-aaf0-4dc4-bef4-25fe2aa58a32 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.280588] env[63531]: DEBUG nova.compute.manager [None req-4b9a0fd3-15d1-440d-90b4-9ce5235e55e0 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 922.280809] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-4b9a0fd3-15d1-440d-90b4-9ce5235e55e0 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 922.282394] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3966d599-ba4b-41d3-88fa-ec637611b889 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.286292] env[63531]: INFO nova.compute.manager [None req-05439101-9bc6-4c60-84fa-463c67262c69 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] instance snapshotting [ 922.288905] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd9f7adb-34a0-454d-b352-e67cd39749b7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.294358] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b9a0fd3-15d1-440d-90b4-9ce5235e55e0 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 922.294659] env[63531]: DEBUG oslo_vmware.api [None req-10dd9342-244a-47e0-b955-67751e3d781a tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Waiting for the task: (returnval){ [ 922.294659] env[63531]: value = "task-1118169" [ 922.294659] env[63531]: _type = "Task" [ 922.294659] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.295135] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7f706ce7-e176-4869-a2dc-0e71323511a9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.312968] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d4f86d3-7909-4c44-bc9b-f7d4c9f5f61c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.319659] env[63531]: DEBUG oslo_vmware.api [None req-4b9a0fd3-15d1-440d-90b4-9ce5235e55e0 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for the task: (returnval){ [ 922.319659] env[63531]: value = "task-1118170" [ 922.319659] env[63531]: _type = "Task" [ 922.319659] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.319977] env[63531]: DEBUG oslo_vmware.api [None req-10dd9342-244a-47e0-b955-67751e3d781a tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118169, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.331746] env[63531]: DEBUG oslo_vmware.api [None req-4b9a0fd3-15d1-440d-90b4-9ce5235e55e0 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118170, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.545839] env[63531]: DEBUG oslo_vmware.api [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5256011c-5694-9dcc-acc1-9c6feaeab8d6, 'name': SearchDatastore_Task, 'duration_secs': 0.010438} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.546149] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.546390] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 922.546630] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.546786] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.546971] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 922.547316] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e0c82d01-bca0-4a8a-86a3-9a8a6a7b249a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.555204] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 922.555271] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 922.556441] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0b598d45-32e6-41b5-884c-0f6b55ddc424 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.560626] env[63531]: DEBUG oslo_vmware.api [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 922.560626] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52e1d73c-995b-ad64-7f7f-b685d59b8b44" [ 922.560626] env[63531]: _type = "Task" [ 922.560626] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.563977] env[63531]: DEBUG nova.network.neutron [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Port e4133e0f-55c1-4b3f-847d-899dd1167eec binding to destination host cpu-1 is already ACTIVE {{(pid=63531) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 922.571998] env[63531]: DEBUG oslo_vmware.api [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52e1d73c-995b-ad64-7f7f-b685d59b8b44, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.577797] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d477826e-3e39-4ca7-bc24-998780166e1c" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.578024] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Processing image d477826e-3e39-4ca7-bc24-998780166e1c {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 922.578259] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d477826e-3e39-4ca7-bc24-998780166e1c/d477826e-3e39-4ca7-bc24-998780166e1c.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.578413] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d477826e-3e39-4ca7-bc24-998780166e1c/d477826e-3e39-4ca7-bc24-998780166e1c.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.578589] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 922.579121] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dd874643-69b4-4419-8488-0978358d0ff5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.586230] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 922.586403] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 922.587088] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c8f56a34-f45b-4127-ac20-5469a554956f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.591465] env[63531]: DEBUG oslo_vmware.api [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Waiting for the task: (returnval){ [ 922.591465] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52664207-be9f-0efa-25a3-f7d3b499c445" [ 922.591465] env[63531]: _type = "Task" [ 922.591465] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.599212] env[63531]: DEBUG oslo_vmware.api [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52664207-be9f-0efa-25a3-f7d3b499c445, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.608051] env[63531]: DEBUG oslo_concurrency.lockutils [None req-935264d7-e19f-4031-a0a0-6f80fb185255 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.866s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.610295] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.734s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.611799] env[63531]: INFO nova.compute.claims [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 922.631232] env[63531]: INFO nova.scheduler.client.report [None req-935264d7-e19f-4031-a0a0-6f80fb185255 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Deleted allocations for instance e95d3273-8216-47cc-95b6-99301366a827 [ 922.781747] env[63531]: DEBUG nova.compute.manager [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 922.784351] env[63531]: DEBUG oslo_concurrency.lockutils [req-7e9e3cc9-43ba-4758-bf87-b5c3a6243411 req-ed25c8d5-44fd-4f04-84d4-f00ba98c41b9 service nova] Releasing lock "refresh_cache-2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.807843] env[63531]: DEBUG oslo_vmware.api [None req-10dd9342-244a-47e0-b955-67751e3d781a tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118169, 'name': ReconfigVM_Task, 'duration_secs': 0.261016} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.811360] env[63531]: DEBUG nova.virt.hardware [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 922.811360] env[63531]: DEBUG nova.virt.hardware [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 922.811360] env[63531]: DEBUG nova.virt.hardware [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 922.811360] env[63531]: DEBUG nova.virt.hardware [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 922.811360] env[63531]: DEBUG nova.virt.hardware [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 922.811360] env[63531]: DEBUG nova.virt.hardware [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 922.811360] env[63531]: DEBUG nova.virt.hardware [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 922.811360] env[63531]: DEBUG nova.virt.hardware [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 922.811665] env[63531]: DEBUG nova.virt.hardware [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 922.811700] env[63531]: DEBUG nova.virt.hardware [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 922.811896] env[63531]: DEBUG nova.virt.hardware [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 922.812227] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-10dd9342-244a-47e0-b955-67751e3d781a tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Reconfigured VM instance instance-0000003c to detach disk 2001 {{(pid=63531) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 922.817243] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c3fccb5-14f5-4eb8-866f-7a618d736c07 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.819695] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-87bced64-2470-453b-b157-fc6224569d69 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.829781] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-05439101-9bc6-4c60-84fa-463c67262c69 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Creating Snapshot of the VM instance {{(pid=63531) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 922.833247] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-a6717d8c-4b66-4aa0-aedf-889e3efaa0e6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.838775] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94c6a982-2fe3-4042-9885-43052b4e4eed {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.847087] env[63531]: DEBUG oslo_vmware.api [None req-4b9a0fd3-15d1-440d-90b4-9ce5235e55e0 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118170, 'name': PowerOffVM_Task, 'duration_secs': 0.192734} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.847428] env[63531]: DEBUG oslo_vmware.api [None req-05439101-9bc6-4c60-84fa-463c67262c69 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Waiting for the task: (returnval){ [ 922.847428] env[63531]: value = "task-1118172" [ 922.847428] env[63531]: _type = "Task" [ 922.847428] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.847781] env[63531]: DEBUG oslo_vmware.api [None req-10dd9342-244a-47e0-b955-67751e3d781a tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Waiting for the task: (returnval){ [ 922.847781] env[63531]: value = "task-1118171" [ 922.847781] env[63531]: _type = "Task" [ 922.847781] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.848402] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b9a0fd3-15d1-440d-90b4-9ce5235e55e0 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 922.848589] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-4b9a0fd3-15d1-440d-90b4-9ce5235e55e0 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 922.848910] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0017585a-c0e2-4deb-b5a4-dea3937429c5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.864364] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Instance VIF info [] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 922.874986] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Creating folder: Project (b24aa128411445199f3e845a9e8b0ed7). Parent ref: group-v244585. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 922.876250] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cf00a870-7e12-444f-80cf-a46ac0a7f0c4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.884138] env[63531]: DEBUG oslo_vmware.api [None req-05439101-9bc6-4c60-84fa-463c67262c69 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1118172, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.884445] env[63531]: DEBUG oslo_vmware.api [None req-10dd9342-244a-47e0-b955-67751e3d781a tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118171, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.895921] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Created folder: Project (b24aa128411445199f3e845a9e8b0ed7) in parent group-v244585. [ 922.896149] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Creating folder: Instances. Parent ref: group-v244747. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 922.896407] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e49e5b62-0184-41cc-9542-e6e5fd7ec853 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.904513] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Created folder: Instances in parent group-v244747. [ 922.904838] env[63531]: DEBUG oslo.service.loopingcall [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 922.905338] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 922.905338] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-995f4553-583d-45fd-bf51-54e44c50e501 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.925866] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 922.925866] env[63531]: value = "task-1118176" [ 922.925866] env[63531]: _type = "Task" [ 922.925866] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.934269] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118176, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.937938] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-4b9a0fd3-15d1-440d-90b4-9ce5235e55e0 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 922.938029] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-4b9a0fd3-15d1-440d-90b4-9ce5235e55e0 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 922.938190] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b9a0fd3-15d1-440d-90b4-9ce5235e55e0 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Deleting the datastore file [datastore1] 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 922.938460] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a17d17e4-bdb6-4aaf-aa8d-446622e73549 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.944709] env[63531]: DEBUG oslo_vmware.api [None req-4b9a0fd3-15d1-440d-90b4-9ce5235e55e0 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for the task: (returnval){ [ 922.944709] env[63531]: value = "task-1118177" [ 922.944709] env[63531]: _type = "Task" [ 922.944709] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.952827] env[63531]: DEBUG oslo_vmware.api [None req-4b9a0fd3-15d1-440d-90b4-9ce5235e55e0 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118177, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.074210] env[63531]: DEBUG oslo_vmware.api [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52e1d73c-995b-ad64-7f7f-b685d59b8b44, 'name': SearchDatastore_Task, 'duration_secs': 0.008288} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.075033] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f1fe5bb3-11e6-41f9-bf26-281a27cb7a4d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.080667] env[63531]: DEBUG oslo_vmware.api [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 923.080667] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52438073-d6c1-d652-4b50-dcf3d9212fac" [ 923.080667] env[63531]: _type = "Task" [ 923.080667] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.088030] env[63531]: DEBUG oslo_vmware.api [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52438073-d6c1-d652-4b50-dcf3d9212fac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.100431] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Preparing fetch location {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 923.100983] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Fetch image to [datastore2] OSTACK_IMG_a108e51e-356f-48b7-9db1-3f8ae81f2b55/OSTACK_IMG_a108e51e-356f-48b7-9db1-3f8ae81f2b55.vmdk {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 923.101340] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Downloading stream optimized image d477826e-3e39-4ca7-bc24-998780166e1c to [datastore2] OSTACK_IMG_a108e51e-356f-48b7-9db1-3f8ae81f2b55/OSTACK_IMG_a108e51e-356f-48b7-9db1-3f8ae81f2b55.vmdk on the data store datastore2 as vApp {{(pid=63531) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 923.101571] env[63531]: DEBUG nova.virt.vmwareapi.images [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Downloading image file data d477826e-3e39-4ca7-bc24-998780166e1c to the ESX as VM named 'OSTACK_IMG_a108e51e-356f-48b7-9db1-3f8ae81f2b55' {{(pid=63531) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 923.149986] env[63531]: DEBUG oslo_concurrency.lockutils [None req-935264d7-e19f-4031-a0a0-6f80fb185255 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "e95d3273-8216-47cc-95b6-99301366a827" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.266s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.190242] env[63531]: DEBUG oslo_vmware.rw_handles [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 923.190242] env[63531]: value = "resgroup-9" [ 923.190242] env[63531]: _type = "ResourcePool" [ 923.190242] env[63531]: }. {{(pid=63531) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 923.194677] env[63531]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-b9bf99c5-f329-4108-a233-af71030ab505 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.215386] env[63531]: DEBUG oslo_vmware.rw_handles [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Lease: (returnval){ [ 923.215386] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52f8a763-6b04-b222-e275-bd3975be28e5" [ 923.215386] env[63531]: _type = "HttpNfcLease" [ 923.215386] env[63531]: } obtained for vApp import into resource pool (val){ [ 923.215386] env[63531]: value = "resgroup-9" [ 923.215386] env[63531]: _type = "ResourcePool" [ 923.215386] env[63531]: }. {{(pid=63531) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 923.217339] env[63531]: DEBUG oslo_vmware.api [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Waiting for the lease: (returnval){ [ 923.217339] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52f8a763-6b04-b222-e275-bd3975be28e5" [ 923.217339] env[63531]: _type = "HttpNfcLease" [ 923.217339] env[63531]: } to be ready. {{(pid=63531) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 923.226026] env[63531]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 923.226026] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52f8a763-6b04-b222-e275-bd3975be28e5" [ 923.226026] env[63531]: _type = "HttpNfcLease" [ 923.226026] env[63531]: } is initializing. {{(pid=63531) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 923.361782] env[63531]: DEBUG oslo_vmware.api [None req-10dd9342-244a-47e0-b955-67751e3d781a tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118171, 'name': ReconfigVM_Task, 'duration_secs': 0.143882} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.365533] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-10dd9342-244a-47e0-b955-67751e3d781a tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244725', 'volume_id': 'c89ce39a-7d3d-468e-b6d6-7df150ed8817', 'name': 'volume-c89ce39a-7d3d-468e-b6d6-7df150ed8817', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f46d8a85-6ab7-436a-bc5e-40d0f80974e5', 'attached_at': '', 'detached_at': '', 'volume_id': 'c89ce39a-7d3d-468e-b6d6-7df150ed8817', 'serial': 'c89ce39a-7d3d-468e-b6d6-7df150ed8817'} {{(pid=63531) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 923.367904] env[63531]: DEBUG oslo_vmware.api [None req-05439101-9bc6-4c60-84fa-463c67262c69 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1118172, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.438577] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118176, 'name': CreateVM_Task, 'duration_secs': 0.358468} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.438841] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 923.439504] env[63531]: DEBUG oslo_concurrency.lockutils [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.439613] env[63531]: DEBUG oslo_concurrency.lockutils [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.440033] env[63531]: DEBUG oslo_concurrency.lockutils [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 923.440367] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7f3b196c-7f08-490a-a878-7cc00fb5974b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.446062] env[63531]: DEBUG oslo_vmware.api [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Waiting for the task: (returnval){ [ 923.446062] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52942578-397f-bacc-fa74-91f212820daf" [ 923.446062] env[63531]: _type = "Task" [ 923.446062] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.460030] env[63531]: DEBUG oslo_vmware.api [None req-4b9a0fd3-15d1-440d-90b4-9ce5235e55e0 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118177, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.234634} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.464614] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b9a0fd3-15d1-440d-90b4-9ce5235e55e0 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 923.464892] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-4b9a0fd3-15d1-440d-90b4-9ce5235e55e0 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 923.465148] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-4b9a0fd3-15d1-440d-90b4-9ce5235e55e0 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 923.465385] env[63531]: INFO nova.compute.manager [None req-4b9a0fd3-15d1-440d-90b4-9ce5235e55e0 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Took 1.18 seconds to destroy the instance on the hypervisor. [ 923.465704] env[63531]: DEBUG oslo.service.loopingcall [None req-4b9a0fd3-15d1-440d-90b4-9ce5235e55e0 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 923.465974] env[63531]: DEBUG oslo_vmware.api [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52942578-397f-bacc-fa74-91f212820daf, 'name': SearchDatastore_Task, 'duration_secs': 0.0093} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.466301] env[63531]: DEBUG nova.compute.manager [-] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 923.466392] env[63531]: DEBUG nova.network.neutron [-] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 923.469132] env[63531]: DEBUG oslo_concurrency.lockutils [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.469132] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 923.469265] env[63531]: DEBUG oslo_concurrency.lockutils [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.594263] env[63531]: DEBUG oslo_concurrency.lockutils [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquiring lock "9a3fad46-a15a-451c-bdab-a3c8cc8add07-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.594530] env[63531]: DEBUG oslo_concurrency.lockutils [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "9a3fad46-a15a-451c-bdab-a3c8cc8add07-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.594719] env[63531]: DEBUG oslo_concurrency.lockutils [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "9a3fad46-a15a-451c-bdab-a3c8cc8add07-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.602371] env[63531]: DEBUG oslo_vmware.api [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52438073-d6c1-d652-4b50-dcf3d9212fac, 'name': SearchDatastore_Task, 'duration_secs': 0.103327} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.603104] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.603104] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] 6118e242-c0db-4707-a5f9-3d12b823935f/6118e242-c0db-4707-a5f9-3d12b823935f.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 923.603286] env[63531]: DEBUG oslo_concurrency.lockutils [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.603468] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 923.603703] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e2415b98-906b-45d6-b8e9-ce39b98bbda6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.606434] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-494c0230-d202-4b8c-8745-5b8267ddc50d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.616040] env[63531]: DEBUG oslo_vmware.api [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 923.616040] env[63531]: value = "task-1118179" [ 923.616040] env[63531]: _type = "Task" [ 923.616040] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.618110] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 923.618110] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 923.621254] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ac6f65ac-d344-49ad-a381-12bddd2b90ed {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.632772] env[63531]: DEBUG oslo_vmware.api [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Waiting for the task: (returnval){ [ 923.632772] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52ef5745-598f-17d8-04e7-e3e22b38e979" [ 923.632772] env[63531]: _type = "Task" [ 923.632772] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.632772] env[63531]: DEBUG oslo_vmware.api [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118179, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.641174] env[63531]: DEBUG oslo_vmware.api [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52ef5745-598f-17d8-04e7-e3e22b38e979, 'name': SearchDatastore_Task, 'duration_secs': 0.009253} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.642013] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e8a6c376-ef6e-4654-b1d5-0ad5f5293e0d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.647905] env[63531]: DEBUG oslo_vmware.api [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Waiting for the task: (returnval){ [ 923.647905] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]529499c8-d840-4c12-8d40-7bc2b706578c" [ 923.647905] env[63531]: _type = "Task" [ 923.647905] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.659278] env[63531]: DEBUG oslo_vmware.api [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]529499c8-d840-4c12-8d40-7bc2b706578c, 'name': SearchDatastore_Task, 'duration_secs': 0.008627} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.659864] env[63531]: DEBUG oslo_concurrency.lockutils [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.660174] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] e065f052-ba3e-4783-8953-5dc200d1f3e9/e065f052-ba3e-4783-8953-5dc200d1f3e9.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 923.660436] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-af6150b7-377a-428f-9ada-b129b5b4b158 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.666692] env[63531]: DEBUG oslo_vmware.api [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Waiting for the task: (returnval){ [ 923.666692] env[63531]: value = "task-1118180" [ 923.666692] env[63531]: _type = "Task" [ 923.666692] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.677881] env[63531]: DEBUG oslo_vmware.api [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Task: {'id': task-1118180, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.722505] env[63531]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 923.722505] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52f8a763-6b04-b222-e275-bd3975be28e5" [ 923.722505] env[63531]: _type = "HttpNfcLease" [ 923.722505] env[63531]: } is initializing. {{(pid=63531) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 923.860430] env[63531]: DEBUG oslo_vmware.api [None req-05439101-9bc6-4c60-84fa-463c67262c69 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1118172, 'name': CreateSnapshot_Task, 'duration_secs': 0.739085} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.860724] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-05439101-9bc6-4c60-84fa-463c67262c69 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Created Snapshot of the VM instance {{(pid=63531) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 923.861473] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a982b612-38c9-4e67-9674-7bdc0ae0e31a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.924661] env[63531]: DEBUG nova.objects.instance [None req-10dd9342-244a-47e0-b955-67751e3d781a tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lazy-loading 'flavor' on Instance uuid f46d8a85-6ab7-436a-bc5e-40d0f80974e5 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 923.955525] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be66526d-6346-4076-ac3a-5a0be125a898 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.965188] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74d03034-5c8a-4927-8d2b-b000f90db10d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.998906] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb2ebb1b-303e-4fc1-b9db-5d7eb9ba25b4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.009494] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80179939-1625-48c3-ad9d-e7ef5bbc035e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.023238] env[63531]: DEBUG nova.compute.provider_tree [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 924.026407] env[63531]: DEBUG nova.compute.manager [req-58e8ed3c-faa2-4351-9b8c-47232754de1c req-04cf8209-cfb5-4df9-827c-c712ccbcc2e3 service nova] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Received event network-vif-deleted-073adca7-1360-4703-9eb1-d86758bb8002 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 924.026407] env[63531]: INFO nova.compute.manager [req-58e8ed3c-faa2-4351-9b8c-47232754de1c req-04cf8209-cfb5-4df9-827c-c712ccbcc2e3 service nova] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Neutron deleted interface 073adca7-1360-4703-9eb1-d86758bb8002; detaching it from the instance and deleting it from the info cache [ 924.026569] env[63531]: DEBUG nova.network.neutron [req-58e8ed3c-faa2-4351-9b8c-47232754de1c req-04cf8209-cfb5-4df9-827c-c712ccbcc2e3 service nova] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.126464] env[63531]: DEBUG oslo_vmware.api [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118179, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.178164] env[63531]: DEBUG oslo_vmware.api [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Task: {'id': task-1118180, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.223991] env[63531]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 924.223991] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52f8a763-6b04-b222-e275-bd3975be28e5" [ 924.223991] env[63531]: _type = "HttpNfcLease" [ 924.223991] env[63531]: } is initializing. {{(pid=63531) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 924.383910] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-05439101-9bc6-4c60-84fa-463c67262c69 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Creating linked-clone VM from snapshot {{(pid=63531) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 924.384744] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-668226dd-adfc-48b7-81ec-8dfcc0ed058e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.396332] env[63531]: DEBUG oslo_vmware.api [None req-05439101-9bc6-4c60-84fa-463c67262c69 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Waiting for the task: (returnval){ [ 924.396332] env[63531]: value = "task-1118181" [ 924.396332] env[63531]: _type = "Task" [ 924.396332] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.408825] env[63531]: DEBUG oslo_vmware.api [None req-05439101-9bc6-4c60-84fa-463c67262c69 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1118181, 'name': CloneVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.448459] env[63531]: DEBUG nova.network.neutron [-] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.529808] env[63531]: DEBUG nova.scheduler.client.report [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 924.536864] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-df11cf0b-39a8-4340-9f7c-1d6451de3f7a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.550670] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c0329d2-ed98-4dd3-ab4f-604093ae526d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.595493] env[63531]: DEBUG nova.compute.manager [req-58e8ed3c-faa2-4351-9b8c-47232754de1c req-04cf8209-cfb5-4df9-827c-c712ccbcc2e3 service nova] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Detach interface failed, port_id=073adca7-1360-4703-9eb1-d86758bb8002, reason: Instance 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9 could not be found. {{(pid=63531) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 924.629638] env[63531]: DEBUG oslo_vmware.api [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118179, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.651962] env[63531]: DEBUG oslo_concurrency.lockutils [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquiring lock "refresh_cache-9a3fad46-a15a-451c-bdab-a3c8cc8add07" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.652271] env[63531]: DEBUG oslo_concurrency.lockutils [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquired lock "refresh_cache-9a3fad46-a15a-451c-bdab-a3c8cc8add07" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.652500] env[63531]: DEBUG nova.network.neutron [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 924.679829] env[63531]: DEBUG oslo_vmware.api [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Task: {'id': task-1118180, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.725251] env[63531]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 924.725251] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52f8a763-6b04-b222-e275-bd3975be28e5" [ 924.725251] env[63531]: _type = "HttpNfcLease" [ 924.725251] env[63531]: } is initializing. {{(pid=63531) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 924.834464] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1aa85d41-e218-4c46-a7dc-85adba5e28f5 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Acquiring lock "f46d8a85-6ab7-436a-bc5e-40d0f80974e5" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.908394] env[63531]: DEBUG oslo_vmware.api [None req-05439101-9bc6-4c60-84fa-463c67262c69 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1118181, 'name': CloneVM_Task} progress is 94%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.933764] env[63531]: DEBUG oslo_concurrency.lockutils [None req-10dd9342-244a-47e0-b955-67751e3d781a tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lock "f46d8a85-6ab7-436a-bc5e-40d0f80974e5" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.300s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.936895] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1aa85d41-e218-4c46-a7dc-85adba5e28f5 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lock "f46d8a85-6ab7-436a-bc5e-40d0f80974e5" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.101s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.936895] env[63531]: DEBUG nova.compute.manager [None req-1aa85d41-e218-4c46-a7dc-85adba5e28f5 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 924.936895] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53372c92-5447-433f-8ff7-bdc8893044f9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.945729] env[63531]: DEBUG nova.compute.manager [None req-1aa85d41-e218-4c46-a7dc-85adba5e28f5 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63531) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 924.947026] env[63531]: DEBUG nova.objects.instance [None req-1aa85d41-e218-4c46-a7dc-85adba5e28f5 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lazy-loading 'flavor' on Instance uuid f46d8a85-6ab7-436a-bc5e-40d0f80974e5 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 924.951644] env[63531]: INFO nova.compute.manager [-] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Took 1.49 seconds to deallocate network for instance. [ 925.039180] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.429s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.039880] env[63531]: DEBUG nova.compute.manager [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 925.042891] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.118s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.043105] env[63531]: DEBUG nova.objects.instance [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Lazy-loading 'resources' on Instance uuid 1ca7b851-2a06-4181-8271-58aafcd322d6 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 925.128718] env[63531]: DEBUG oslo_vmware.api [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118179, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.29713} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.129075] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] 6118e242-c0db-4707-a5f9-3d12b823935f/6118e242-c0db-4707-a5f9-3d12b823935f.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 925.129316] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 925.129586] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7f362afc-751a-467a-91a5-2b24b24c55e4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.136308] env[63531]: DEBUG oslo_vmware.api [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 925.136308] env[63531]: value = "task-1118182" [ 925.136308] env[63531]: _type = "Task" [ 925.136308] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.145128] env[63531]: DEBUG oslo_vmware.api [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118182, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.179197] env[63531]: DEBUG oslo_vmware.api [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Task: {'id': task-1118180, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.252423} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.179513] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] e065f052-ba3e-4783-8953-5dc200d1f3e9/e065f052-ba3e-4783-8953-5dc200d1f3e9.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 925.179740] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 925.180031] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-033f2c6a-7154-4c21-b54f-c662245a77a2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.186486] env[63531]: DEBUG oslo_vmware.api [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Waiting for the task: (returnval){ [ 925.186486] env[63531]: value = "task-1118183" [ 925.186486] env[63531]: _type = "Task" [ 925.186486] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.195564] env[63531]: DEBUG oslo_vmware.api [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Task: {'id': task-1118183, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.225426] env[63531]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 925.225426] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52f8a763-6b04-b222-e275-bd3975be28e5" [ 925.225426] env[63531]: _type = "HttpNfcLease" [ 925.225426] env[63531]: } is initializing. {{(pid=63531) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 925.410497] env[63531]: DEBUG oslo_vmware.api [None req-05439101-9bc6-4c60-84fa-463c67262c69 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1118181, 'name': CloneVM_Task} progress is 94%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.442921] env[63531]: DEBUG nova.network.neutron [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Updating instance_info_cache with network_info: [{"id": "e4133e0f-55c1-4b3f-847d-899dd1167eec", "address": "fa:16:3e:90:40:9b", "network": {"id": "c4f4c994-36ab-42ab-94f8-78f29cbafa50", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-994507698-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "57e9bf4ee0f1463fa67b0f8deff30fdc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "767a3a48-41d4-4a0c-961d-0024837f63bd", "external-id": "nsx-vlan-transportzone-141", "segmentation_id": 141, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4133e0f-55", "ovs_interfaceid": "e4133e0f-55c1-4b3f-847d-899dd1167eec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 925.454803] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-1aa85d41-e218-4c46-a7dc-85adba5e28f5 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 925.455028] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-31158bee-bea7-495d-9da9-c22764faf783 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.459389] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4b9a0fd3-15d1-440d-90b4-9ce5235e55e0 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.462803] env[63531]: DEBUG oslo_vmware.api [None req-1aa85d41-e218-4c46-a7dc-85adba5e28f5 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Waiting for the task: (returnval){ [ 925.462803] env[63531]: value = "task-1118184" [ 925.462803] env[63531]: _type = "Task" [ 925.462803] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.471949] env[63531]: DEBUG oslo_vmware.api [None req-1aa85d41-e218-4c46-a7dc-85adba5e28f5 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118184, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.547036] env[63531]: DEBUG nova.compute.utils [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 925.548730] env[63531]: DEBUG nova.objects.instance [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Lazy-loading 'numa_topology' on Instance uuid 1ca7b851-2a06-4181-8271-58aafcd322d6 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 925.550080] env[63531]: DEBUG nova.compute.manager [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 925.550179] env[63531]: DEBUG nova.network.neutron [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 925.594316] env[63531]: DEBUG nova.policy [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7c5a958ed3574aff9c1a6a144b42718d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f71c7db0a3614995becc25d9a8dce787', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 925.646731] env[63531]: DEBUG oslo_vmware.api [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118182, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.084726} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.647039] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 925.647878] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a137d76-347b-4696-b96c-7206b51c59f2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.669864] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Reconfiguring VM instance instance-0000004d to attach disk [datastore2] 6118e242-c0db-4707-a5f9-3d12b823935f/6118e242-c0db-4707-a5f9-3d12b823935f.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 925.670189] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-61583ce2-18dc-48cb-952f-20e8ff3aa42d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.692245] env[63531]: DEBUG oslo_vmware.api [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 925.692245] env[63531]: value = "task-1118185" [ 925.692245] env[63531]: _type = "Task" [ 925.692245] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.700070] env[63531]: DEBUG oslo_vmware.api [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Task: {'id': task-1118183, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.098639} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.700070] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 925.700683] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf4acb19-02ad-493b-b491-7e57001c1019 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.706368] env[63531]: DEBUG oslo_vmware.api [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118185, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.724434] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Reconfiguring VM instance instance-00000050 to attach disk [datastore2] e065f052-ba3e-4783-8953-5dc200d1f3e9/e065f052-ba3e-4783-8953-5dc200d1f3e9.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 925.724829] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b0faf347-d15c-48f0-b5a5-cc26b28e846c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.746592] env[63531]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 925.746592] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52f8a763-6b04-b222-e275-bd3975be28e5" [ 925.746592] env[63531]: _type = "HttpNfcLease" [ 925.746592] env[63531]: } is initializing. {{(pid=63531) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 925.748163] env[63531]: DEBUG oslo_vmware.api [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Waiting for the task: (returnval){ [ 925.748163] env[63531]: value = "task-1118186" [ 925.748163] env[63531]: _type = "Task" [ 925.748163] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.756419] env[63531]: DEBUG oslo_vmware.api [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Task: {'id': task-1118186, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.892146] env[63531]: DEBUG nova.network.neutron [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Successfully created port: 7433ce46-ff5a-4e78-bcc9-9dcf2264a01b {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 925.912930] env[63531]: DEBUG oslo_vmware.api [None req-05439101-9bc6-4c60-84fa-463c67262c69 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1118181, 'name': CloneVM_Task} progress is 94%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.945836] env[63531]: DEBUG oslo_concurrency.lockutils [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Releasing lock "refresh_cache-9a3fad46-a15a-451c-bdab-a3c8cc8add07" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.974996] env[63531]: DEBUG oslo_vmware.api [None req-1aa85d41-e218-4c46-a7dc-85adba5e28f5 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118184, 'name': PowerOffVM_Task, 'duration_secs': 0.357403} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.975351] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-1aa85d41-e218-4c46-a7dc-85adba5e28f5 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 925.975780] env[63531]: DEBUG nova.compute.manager [None req-1aa85d41-e218-4c46-a7dc-85adba5e28f5 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 925.976633] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d78e534e-3142-483a-aa4a-9c6ae19e6ff6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.002017] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquiring lock "911f23f7-e320-48f3-87a9-0239b013feff" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.002369] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "911f23f7-e320-48f3-87a9-0239b013feff" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.050495] env[63531]: DEBUG nova.compute.manager [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 926.054175] env[63531]: DEBUG nova.objects.base [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Object Instance<1ca7b851-2a06-4181-8271-58aafcd322d6> lazy-loaded attributes: resources,numa_topology {{(pid=63531) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 926.203857] env[63531]: DEBUG oslo_vmware.api [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118185, 'name': ReconfigVM_Task, 'duration_secs': 0.381336} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.203949] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Reconfigured VM instance instance-0000004d to attach disk [datastore2] 6118e242-c0db-4707-a5f9-3d12b823935f/6118e242-c0db-4707-a5f9-3d12b823935f.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 926.204591] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7553b8ad-9ed9-4d04-ab50-7bd7481583bd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.211773] env[63531]: DEBUG oslo_vmware.api [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 926.211773] env[63531]: value = "task-1118187" [ 926.211773] env[63531]: _type = "Task" [ 926.211773] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.229040] env[63531]: DEBUG oslo_vmware.api [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118187, 'name': Rename_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.235903] env[63531]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 926.235903] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52f8a763-6b04-b222-e275-bd3975be28e5" [ 926.235903] env[63531]: _type = "HttpNfcLease" [ 926.235903] env[63531]: } is initializing. {{(pid=63531) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 926.259901] env[63531]: DEBUG oslo_vmware.api [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Task: {'id': task-1118186, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.366458] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e543ce6d-d0db-4c21-bf1f-f725e978ee23 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.373698] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbc7dbbd-ab70-4c85-b12f-b8eb136c72f8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.411785] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-206ba67c-4d23-41af-8fe1-5263afcc7488 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.419827] env[63531]: DEBUG oslo_vmware.api [None req-05439101-9bc6-4c60-84fa-463c67262c69 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1118181, 'name': CloneVM_Task} progress is 100%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.423328] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4618f2bd-58b5-4d9f-a624-c274945ca5cc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.441590] env[63531]: DEBUG nova.compute.provider_tree [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 926.475060] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4107fbc1-3e0f-4c9c-a292-7ce6bc2c6255 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.494050] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d5ed37e-acd3-445c-b0b0-b13b79ac7e57 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.497233] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1aa85d41-e218-4c46-a7dc-85adba5e28f5 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lock "f46d8a85-6ab7-436a-bc5e-40d0f80974e5" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.562s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.502424] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Updating instance '9a3fad46-a15a-451c-bdab-a3c8cc8add07' progress to 83 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 926.508349] env[63531]: DEBUG nova.compute.manager [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 926.721677] env[63531]: DEBUG oslo_vmware.api [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118187, 'name': Rename_Task, 'duration_secs': 0.137683} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.722008] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 926.722264] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b0f230bf-b070-4faa-89c5-dcc5c2da9d45 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.728603] env[63531]: DEBUG oslo_vmware.api [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 926.728603] env[63531]: value = "task-1118188" [ 926.728603] env[63531]: _type = "Task" [ 926.728603] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.739184] env[63531]: DEBUG oslo_vmware.api [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118188, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.740769] env[63531]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 926.740769] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52f8a763-6b04-b222-e275-bd3975be28e5" [ 926.740769] env[63531]: _type = "HttpNfcLease" [ 926.740769] env[63531]: } is ready. {{(pid=63531) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 926.741071] env[63531]: DEBUG oslo_vmware.rw_handles [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 926.741071] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52f8a763-6b04-b222-e275-bd3975be28e5" [ 926.741071] env[63531]: _type = "HttpNfcLease" [ 926.741071] env[63531]: }. {{(pid=63531) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 926.741820] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aade7b8c-ed06-4c7a-8071-e4c12569450e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.748772] env[63531]: DEBUG oslo_vmware.rw_handles [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ef2fe4-b872-a407-d782-f32bfc6da7cd/disk-0.vmdk from lease info. {{(pid=63531) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 926.748965] env[63531]: DEBUG oslo_vmware.rw_handles [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Creating HTTP connection to write to file with size = 31666688 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ef2fe4-b872-a407-d782-f32bfc6da7cd/disk-0.vmdk. {{(pid=63531) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 926.811156] env[63531]: DEBUG oslo_vmware.api [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Task: {'id': task-1118186, 'name': ReconfigVM_Task, 'duration_secs': 0.722197} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.812536] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Reconfigured VM instance instance-00000050 to attach disk [datastore2] e065f052-ba3e-4783-8953-5dc200d1f3e9/e065f052-ba3e-4783-8953-5dc200d1f3e9.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 926.815906] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a0bccf66-3621-40ba-a786-ff2eed15d214 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.819295] env[63531]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-ffaede31-0e32-4b4f-b97e-5a54a6ecee87 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.825979] env[63531]: DEBUG oslo_vmware.api [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Waiting for the task: (returnval){ [ 926.825979] env[63531]: value = "task-1118189" [ 926.825979] env[63531]: _type = "Task" [ 926.825979] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.833699] env[63531]: DEBUG oslo_vmware.api [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Task: {'id': task-1118189, 'name': Rename_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.919413] env[63531]: DEBUG oslo_vmware.api [None req-05439101-9bc6-4c60-84fa-463c67262c69 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1118181, 'name': CloneVM_Task, 'duration_secs': 2.077611} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.919682] env[63531]: INFO nova.virt.vmwareapi.vmops [None req-05439101-9bc6-4c60-84fa-463c67262c69 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Created linked-clone VM from snapshot [ 926.920451] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10ad85c5-3f8f-4764-976a-93a56ed73abc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.928458] env[63531]: DEBUG nova.virt.vmwareapi.images [None req-05439101-9bc6-4c60-84fa-463c67262c69 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Uploading image a2ff4559-1f63-4780-aadb-9407f9d2e977 {{(pid=63531) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 926.942911] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-05439101-9bc6-4c60-84fa-463c67262c69 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Destroying the VM {{(pid=63531) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 926.943267] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-b34afdbc-0fe9-4452-b9c2-f1ccee32eae6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.949997] env[63531]: DEBUG nova.scheduler.client.report [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 926.955044] env[63531]: DEBUG oslo_vmware.api [None req-05439101-9bc6-4c60-84fa-463c67262c69 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Waiting for the task: (returnval){ [ 926.955044] env[63531]: value = "task-1118190" [ 926.955044] env[63531]: _type = "Task" [ 926.955044] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.967163] env[63531]: DEBUG oslo_vmware.api [None req-05439101-9bc6-4c60-84fa-463c67262c69 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1118190, 'name': Destroy_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.008594] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-27fbe191-8068-4544-8bd3-e5f2b4820b94 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Updating instance '9a3fad46-a15a-451c-bdab-a3c8cc8add07' progress to 100 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 927.035366] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.060832] env[63531]: DEBUG nova.compute.manager [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 927.088173] env[63531]: DEBUG nova.virt.hardware [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 927.088466] env[63531]: DEBUG nova.virt.hardware [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 927.088692] env[63531]: DEBUG nova.virt.hardware [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 927.088877] env[63531]: DEBUG nova.virt.hardware [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 927.089086] env[63531]: DEBUG nova.virt.hardware [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 927.089260] env[63531]: DEBUG nova.virt.hardware [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 927.089476] env[63531]: DEBUG nova.virt.hardware [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 927.089642] env[63531]: DEBUG nova.virt.hardware [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 927.089857] env[63531]: DEBUG nova.virt.hardware [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 927.090080] env[63531]: DEBUG nova.virt.hardware [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 927.090294] env[63531]: DEBUG nova.virt.hardware [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 927.091321] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b52f38a-591e-486d-b032-d528e7fed4e1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.103374] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b66f897-2f12-4ff3-a420-b7f260a86339 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.241014] env[63531]: DEBUG oslo_vmware.api [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118188, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.338124] env[63531]: DEBUG oslo_vmware.api [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Task: {'id': task-1118189, 'name': Rename_Task, 'duration_secs': 0.193279} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.340065] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 927.340411] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a4dec05e-fe1a-40f7-9fd1-c449ad95c1f3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.347929] env[63531]: DEBUG oslo_vmware.api [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Waiting for the task: (returnval){ [ 927.347929] env[63531]: value = "task-1118191" [ 927.347929] env[63531]: _type = "Task" [ 927.347929] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.356446] env[63531]: DEBUG nova.compute.manager [req-77ba8f82-e8cf-43f4-a941-57a64753cc0f req-f6c48f8d-3889-4e56-97eb-107e768493f1 service nova] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Received event network-vif-plugged-7433ce46-ff5a-4e78-bcc9-9dcf2264a01b {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 927.356812] env[63531]: DEBUG oslo_concurrency.lockutils [req-77ba8f82-e8cf-43f4-a941-57a64753cc0f req-f6c48f8d-3889-4e56-97eb-107e768493f1 service nova] Acquiring lock "ccf00b09-29fe-4a76-a8af-97774acd77de-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.357114] env[63531]: DEBUG oslo_concurrency.lockutils [req-77ba8f82-e8cf-43f4-a941-57a64753cc0f req-f6c48f8d-3889-4e56-97eb-107e768493f1 service nova] Lock "ccf00b09-29fe-4a76-a8af-97774acd77de-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.357498] env[63531]: DEBUG oslo_concurrency.lockutils [req-77ba8f82-e8cf-43f4-a941-57a64753cc0f req-f6c48f8d-3889-4e56-97eb-107e768493f1 service nova] Lock "ccf00b09-29fe-4a76-a8af-97774acd77de-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.357752] env[63531]: DEBUG nova.compute.manager [req-77ba8f82-e8cf-43f4-a941-57a64753cc0f req-f6c48f8d-3889-4e56-97eb-107e768493f1 service nova] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] No waiting events found dispatching network-vif-plugged-7433ce46-ff5a-4e78-bcc9-9dcf2264a01b {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 927.357969] env[63531]: WARNING nova.compute.manager [req-77ba8f82-e8cf-43f4-a941-57a64753cc0f req-f6c48f8d-3889-4e56-97eb-107e768493f1 service nova] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Received unexpected event network-vif-plugged-7433ce46-ff5a-4e78-bcc9-9dcf2264a01b for instance with vm_state building and task_state spawning. [ 927.363735] env[63531]: DEBUG oslo_vmware.api [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Task: {'id': task-1118191, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.458079] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.414s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.464023] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 21.472s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.464023] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.464023] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63531) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 927.464023] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.270s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.465231] env[63531]: INFO nova.compute.claims [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3fb0499d-8074-4e63-be9f-380730416cc4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 927.480722] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bceeb7d2-2cef-4f7e-a1a8-d958c99bd6ed {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.482698] env[63531]: DEBUG nova.objects.instance [None req-132ff447-284d-4f08-bf28-5afb769e38dd tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lazy-loading 'flavor' on Instance uuid f46d8a85-6ab7-436a-bc5e-40d0f80974e5 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 927.490786] env[63531]: DEBUG oslo_vmware.api [None req-05439101-9bc6-4c60-84fa-463c67262c69 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1118190, 'name': Destroy_Task} progress is 33%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.495965] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2edb9edc-0957-4238-b073-8a47aa4996ac {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.524242] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e08b3725-1dac-413c-9bef-bcc06b0f0caa {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.533869] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fea71e37-8e53-43f8-bbbb-20282df9835d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.568270] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179986MB free_disk=169GB free_vcpus=48 pci_devices=None {{(pid=63531) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 927.568580] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.741496] env[63531]: DEBUG oslo_vmware.api [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118188, 'name': PowerOnVM_Task, 'duration_secs': 0.588353} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.741903] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 927.742183] env[63531]: DEBUG nova.compute.manager [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 927.744383] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0921d166-dacd-4a5c-a36a-2cf8549a5d4c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.860384] env[63531]: DEBUG oslo_vmware.api [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Task: {'id': task-1118191, 'name': PowerOnVM_Task} progress is 90%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.950265] env[63531]: DEBUG nova.network.neutron [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Successfully updated port: 7433ce46-ff5a-4e78-bcc9-9dcf2264a01b {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 927.975131] env[63531]: DEBUG oslo_vmware.api [None req-05439101-9bc6-4c60-84fa-463c67262c69 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1118190, 'name': Destroy_Task, 'duration_secs': 0.623242} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.977114] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-05439101-9bc6-4c60-84fa-463c67262c69 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Destroyed the VM [ 927.977314] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-05439101-9bc6-4c60-84fa-463c67262c69 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Deleting Snapshot of the VM instance {{(pid=63531) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 927.977915] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-fd5ff103-d2b1-469b-89b4-7df8797e6e62 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.989900] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e3be4c43-80b6-454a-8012-a555c506f6d6 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Lock "1ca7b851-2a06-4181-8271-58aafcd322d6" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 47.999s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.991325] env[63531]: DEBUG oslo_vmware.api [None req-05439101-9bc6-4c60-84fa-463c67262c69 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Waiting for the task: (returnval){ [ 927.991325] env[63531]: value = "task-1118192" [ 927.991325] env[63531]: _type = "Task" [ 927.991325] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.991613] env[63531]: DEBUG oslo_concurrency.lockutils [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Lock "1ca7b851-2a06-4181-8271-58aafcd322d6" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 21.447s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.991800] env[63531]: INFO nova.compute.manager [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Unshelving [ 927.995475] env[63531]: DEBUG oslo_concurrency.lockutils [None req-132ff447-284d-4f08-bf28-5afb769e38dd tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Acquiring lock "refresh_cache-f46d8a85-6ab7-436a-bc5e-40d0f80974e5" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.995475] env[63531]: DEBUG oslo_concurrency.lockutils [None req-132ff447-284d-4f08-bf28-5afb769e38dd tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Acquired lock "refresh_cache-f46d8a85-6ab7-436a-bc5e-40d0f80974e5" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.995604] env[63531]: DEBUG nova.network.neutron [None req-132ff447-284d-4f08-bf28-5afb769e38dd tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 927.995891] env[63531]: DEBUG nova.objects.instance [None req-132ff447-284d-4f08-bf28-5afb769e38dd tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lazy-loading 'info_cache' on Instance uuid f46d8a85-6ab7-436a-bc5e-40d0f80974e5 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 928.010445] env[63531]: DEBUG oslo_vmware.api [None req-05439101-9bc6-4c60-84fa-463c67262c69 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1118192, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.052351] env[63531]: DEBUG nova.compute.manager [req-7ed739dd-2efc-4d16-bf97-b1b496c898ff req-2b9e801e-3938-4ce6-abcd-d3fb3b5a1516 service nova] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Received event network-changed-7433ce46-ff5a-4e78-bcc9-9dcf2264a01b {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 928.052605] env[63531]: DEBUG nova.compute.manager [req-7ed739dd-2efc-4d16-bf97-b1b496c898ff req-2b9e801e-3938-4ce6-abcd-d3fb3b5a1516 service nova] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Refreshing instance network info cache due to event network-changed-7433ce46-ff5a-4e78-bcc9-9dcf2264a01b. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 928.052877] env[63531]: DEBUG oslo_concurrency.lockutils [req-7ed739dd-2efc-4d16-bf97-b1b496c898ff req-2b9e801e-3938-4ce6-abcd-d3fb3b5a1516 service nova] Acquiring lock "refresh_cache-ccf00b09-29fe-4a76-a8af-97774acd77de" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 928.053089] env[63531]: DEBUG oslo_concurrency.lockutils [req-7ed739dd-2efc-4d16-bf97-b1b496c898ff req-2b9e801e-3938-4ce6-abcd-d3fb3b5a1516 service nova] Acquired lock "refresh_cache-ccf00b09-29fe-4a76-a8af-97774acd77de" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.053288] env[63531]: DEBUG nova.network.neutron [req-7ed739dd-2efc-4d16-bf97-b1b496c898ff req-2b9e801e-3938-4ce6-abcd-d3fb3b5a1516 service nova] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Refreshing network info cache for port 7433ce46-ff5a-4e78-bcc9-9dcf2264a01b {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 928.113835] env[63531]: DEBUG oslo_vmware.rw_handles [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Completed reading data from the image iterator. {{(pid=63531) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 928.113835] env[63531]: DEBUG oslo_vmware.rw_handles [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ef2fe4-b872-a407-d782-f32bfc6da7cd/disk-0.vmdk. {{(pid=63531) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 928.115296] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae14d02b-e876-4773-b333-f4f486a2195f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.121797] env[63531]: DEBUG oslo_vmware.rw_handles [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ef2fe4-b872-a407-d782-f32bfc6da7cd/disk-0.vmdk is in state: ready. {{(pid=63531) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 928.121983] env[63531]: DEBUG oslo_vmware.rw_handles [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ef2fe4-b872-a407-d782-f32bfc6da7cd/disk-0.vmdk. {{(pid=63531) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 928.122208] env[63531]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-5843f21b-2ad1-40a7-bc92-3a7a20d45668 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.264873] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.360900] env[63531]: DEBUG oslo_vmware.api [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Task: {'id': task-1118191, 'name': PowerOnVM_Task, 'duration_secs': 0.55078} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.361583] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 928.361803] env[63531]: INFO nova.compute.manager [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Took 5.58 seconds to spawn the instance on the hypervisor. [ 928.361986] env[63531]: DEBUG nova.compute.manager [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 928.362823] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66213cc2-a89f-44e6-9e82-3f6acaef5659 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.452984] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquiring lock "refresh_cache-ccf00b09-29fe-4a76-a8af-97774acd77de" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 928.464342] env[63531]: DEBUG oslo_vmware.rw_handles [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ef2fe4-b872-a407-d782-f32bfc6da7cd/disk-0.vmdk. {{(pid=63531) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 928.464342] env[63531]: INFO nova.virt.vmwareapi.images [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Downloaded image file data d477826e-3e39-4ca7-bc24-998780166e1c [ 928.464607] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1178c4a6-7bf6-437b-aa77-aa7174cf28dc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.491245] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fc2111d5-9f33-4d73-83d3-6f9436b4044d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.502971] env[63531]: DEBUG oslo_vmware.api [None req-05439101-9bc6-4c60-84fa-463c67262c69 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1118192, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.506719] env[63531]: DEBUG nova.objects.base [None req-132ff447-284d-4f08-bf28-5afb769e38dd tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=63531) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 928.517121] env[63531]: INFO nova.virt.vmwareapi.images [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] The imported VM was unregistered [ 928.519064] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Caching image {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 928.519297] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Creating directory with path [datastore2] devstack-image-cache_base/d477826e-3e39-4ca7-bc24-998780166e1c {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 928.520161] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3b9219a9-2b91-4912-b111-ce8a267f2dfb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.546360] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Created directory with path [datastore2] devstack-image-cache_base/d477826e-3e39-4ca7-bc24-998780166e1c {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 928.546417] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_a108e51e-356f-48b7-9db1-3f8ae81f2b55/OSTACK_IMG_a108e51e-356f-48b7-9db1-3f8ae81f2b55.vmdk to [datastore2] devstack-image-cache_base/d477826e-3e39-4ca7-bc24-998780166e1c/d477826e-3e39-4ca7-bc24-998780166e1c.vmdk. {{(pid=63531) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 928.548687] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-7584ebf4-a0be-41e4-8c8e-15305853a231 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.561844] env[63531]: DEBUG oslo_vmware.api [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Waiting for the task: (returnval){ [ 928.561844] env[63531]: value = "task-1118194" [ 928.561844] env[63531]: _type = "Task" [ 928.561844] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.577657] env[63531]: DEBUG oslo_vmware.api [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': task-1118194, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.602131] env[63531]: DEBUG nova.network.neutron [req-7ed739dd-2efc-4d16-bf97-b1b496c898ff req-2b9e801e-3938-4ce6-abcd-d3fb3b5a1516 service nova] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 928.698493] env[63531]: DEBUG nova.network.neutron [req-7ed739dd-2efc-4d16-bf97-b1b496c898ff req-2b9e801e-3938-4ce6-abcd-d3fb3b5a1516 service nova] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 928.819548] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b81a74ca-45e1-4449-b718-ba75af0921e3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.827775] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80cebd3f-4f9e-467f-92ea-f82a17a59677 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.860109] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6aad8ce-f1b6-4c50-89bc-f44911f773ed {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.867805] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8838e97c-0b67-4693-ae6f-b54144f5f10c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.885132] env[63531]: DEBUG nova.compute.provider_tree [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 928.890726] env[63531]: INFO nova.compute.manager [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Took 29.79 seconds to build instance. [ 929.002801] env[63531]: DEBUG oslo_vmware.api [None req-05439101-9bc6-4c60-84fa-463c67262c69 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1118192, 'name': RemoveSnapshot_Task, 'duration_secs': 0.876415} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.003067] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-05439101-9bc6-4c60-84fa-463c67262c69 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Deleted Snapshot of the VM instance {{(pid=63531) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 929.024871] env[63531]: DEBUG oslo_concurrency.lockutils [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.072560] env[63531]: DEBUG oslo_vmware.api [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': task-1118194, 'name': MoveVirtualDisk_Task} progress is 9%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.200132] env[63531]: DEBUG oslo_concurrency.lockutils [req-7ed739dd-2efc-4d16-bf97-b1b496c898ff req-2b9e801e-3938-4ce6-abcd-d3fb3b5a1516 service nova] Releasing lock "refresh_cache-ccf00b09-29fe-4a76-a8af-97774acd77de" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 929.202818] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquired lock "refresh_cache-ccf00b09-29fe-4a76-a8af-97774acd77de" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.202988] env[63531]: DEBUG nova.network.neutron [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 929.240281] env[63531]: DEBUG nova.network.neutron [None req-132ff447-284d-4f08-bf28-5afb769e38dd tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Updating instance_info_cache with network_info: [{"id": "a3fd640b-c47b-4501-ac04-f90f6af8b076", "address": "fa:16:3e:27:ca:51", "network": {"id": "4dd26082-3411-4d99-a78b-c76c15703506", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1190890625-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.196", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ee5fd84e1da04fd984f81578e4085850", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6046aec4-feda-4ef9-bf4a-800de1e0cd3b", "external-id": "nsx-vlan-transportzone-903", "segmentation_id": 903, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3fd640b-c4", "ovs_interfaceid": "a3fd640b-c47b-4501-ac04-f90f6af8b076", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 929.309012] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b9944abd-dda8-40ee-b4da-5c9bb59d9201 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "6118e242-c0db-4707-a5f9-3d12b823935f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.309399] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b9944abd-dda8-40ee-b4da-5c9bb59d9201 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "6118e242-c0db-4707-a5f9-3d12b823935f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.309642] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b9944abd-dda8-40ee-b4da-5c9bb59d9201 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "6118e242-c0db-4707-a5f9-3d12b823935f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.309883] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b9944abd-dda8-40ee-b4da-5c9bb59d9201 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "6118e242-c0db-4707-a5f9-3d12b823935f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.310226] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b9944abd-dda8-40ee-b4da-5c9bb59d9201 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "6118e242-c0db-4707-a5f9-3d12b823935f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.313268] env[63531]: INFO nova.compute.manager [None req-b9944abd-dda8-40ee-b4da-5c9bb59d9201 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Terminating instance [ 929.316977] env[63531]: DEBUG nova.compute.manager [None req-b9944abd-dda8-40ee-b4da-5c9bb59d9201 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 929.317195] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-b9944abd-dda8-40ee-b4da-5c9bb59d9201 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 929.318751] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bda4103-0ff9-4d2a-8ba5-7de66d3c2f16 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.333362] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9944abd-dda8-40ee-b4da-5c9bb59d9201 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 929.333762] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7130b6d1-cb73-42bc-89ed-e21b3a316a02 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.342043] env[63531]: DEBUG oslo_vmware.api [None req-b9944abd-dda8-40ee-b4da-5c9bb59d9201 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 929.342043] env[63531]: value = "task-1118195" [ 929.342043] env[63531]: _type = "Task" [ 929.342043] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.352073] env[63531]: DEBUG oslo_vmware.api [None req-b9944abd-dda8-40ee-b4da-5c9bb59d9201 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118195, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.389751] env[63531]: DEBUG nova.scheduler.client.report [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 929.393524] env[63531]: DEBUG oslo_concurrency.lockutils [None req-df4d93b4-b012-449e-8582-b7b07f07d5bc tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Lock "e065f052-ba3e-4783-8953-5dc200d1f3e9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.300s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.463050] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ff3c91f0-14af-4f9f-8d99-3fe699fd3d7c tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquiring lock "9a3fad46-a15a-451c-bdab-a3c8cc8add07" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.463273] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ff3c91f0-14af-4f9f-8d99-3fe699fd3d7c tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "9a3fad46-a15a-451c-bdab-a3c8cc8add07" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.463524] env[63531]: DEBUG nova.compute.manager [None req-ff3c91f0-14af-4f9f-8d99-3fe699fd3d7c tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Going to confirm migration 2 {{(pid=63531) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 929.508502] env[63531]: WARNING nova.compute.manager [None req-05439101-9bc6-4c60-84fa-463c67262c69 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Image not found during snapshot: nova.exception.ImageNotFound: Image a2ff4559-1f63-4780-aadb-9407f9d2e977 could not be found. [ 929.573858] env[63531]: DEBUG oslo_vmware.api [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': task-1118194, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.736855] env[63531]: DEBUG nova.network.neutron [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 929.743464] env[63531]: DEBUG oslo_concurrency.lockutils [None req-132ff447-284d-4f08-bf28-5afb769e38dd tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Releasing lock "refresh_cache-f46d8a85-6ab7-436a-bc5e-40d0f80974e5" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 929.829628] env[63531]: INFO nova.compute.manager [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Rebuilding instance [ 929.861763] env[63531]: DEBUG oslo_vmware.api [None req-b9944abd-dda8-40ee-b4da-5c9bb59d9201 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118195, 'name': PowerOffVM_Task, 'duration_secs': 0.380407} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.866181] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9944abd-dda8-40ee-b4da-5c9bb59d9201 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 929.866395] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-b9944abd-dda8-40ee-b4da-5c9bb59d9201 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 929.866993] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-64909940-a2c7-4617-8836-d20f691ef265 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.886337] env[63531]: DEBUG nova.compute.manager [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 929.887264] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-174ac5af-5ad9-4bbf-a37c-053e5c6b365c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.900031] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.436s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.900031] env[63531]: DEBUG nova.compute.manager [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3fb0499d-8074-4e63-be9f-380730416cc4] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 929.901344] env[63531]: DEBUG nova.network.neutron [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Updating instance_info_cache with network_info: [{"id": "7433ce46-ff5a-4e78-bcc9-9dcf2264a01b", "address": "fa:16:3e:ad:77:b5", "network": {"id": "913045d9-e90b-4366-987c-7706b73a6147", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1696510389-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f71c7db0a3614995becc25d9a8dce787", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c297fe21-cd0b-4226-813b-a65d2358d034", "external-id": "nsx-vlan-transportzone-98", "segmentation_id": 98, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7433ce46-ff", "ovs_interfaceid": "7433ce46-ff5a-4e78-bcc9-9dcf2264a01b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 929.902797] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 21.129s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.903021] env[63531]: DEBUG nova.objects.instance [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63531) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 929.945787] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-b9944abd-dda8-40ee-b4da-5c9bb59d9201 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 929.948818] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-b9944abd-dda8-40ee-b4da-5c9bb59d9201 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Deleting contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 929.948818] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9944abd-dda8-40ee-b4da-5c9bb59d9201 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Deleting the datastore file [datastore2] 6118e242-c0db-4707-a5f9-3d12b823935f {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 929.948818] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-822a2711-7005-4f53-a9be-eab31ee842da {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.958202] env[63531]: DEBUG oslo_vmware.api [None req-b9944abd-dda8-40ee-b4da-5c9bb59d9201 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 929.958202] env[63531]: value = "task-1118197" [ 929.958202] env[63531]: _type = "Task" [ 929.958202] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.971241] env[63531]: DEBUG oslo_vmware.api [None req-b9944abd-dda8-40ee-b4da-5c9bb59d9201 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118197, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.061972] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ff3c91f0-14af-4f9f-8d99-3fe699fd3d7c tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquiring lock "refresh_cache-9a3fad46-a15a-451c-bdab-a3c8cc8add07" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 930.061972] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ff3c91f0-14af-4f9f-8d99-3fe699fd3d7c tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquired lock "refresh_cache-9a3fad46-a15a-451c-bdab-a3c8cc8add07" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.061972] env[63531]: DEBUG nova.network.neutron [None req-ff3c91f0-14af-4f9f-8d99-3fe699fd3d7c tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 930.061972] env[63531]: DEBUG nova.objects.instance [None req-ff3c91f0-14af-4f9f-8d99-3fe699fd3d7c tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lazy-loading 'info_cache' on Instance uuid 9a3fad46-a15a-451c-bdab-a3c8cc8add07 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 930.074507] env[63531]: DEBUG oslo_vmware.api [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': task-1118194, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.099457] env[63531]: DEBUG oslo_concurrency.lockutils [None req-141fc11b-5904-4c16-af8a-e3b6485bc843 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Acquiring lock "4b5e5821-9def-4b53-be19-355d9e5f81ec" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.099743] env[63531]: DEBUG oslo_concurrency.lockutils [None req-141fc11b-5904-4c16-af8a-e3b6485bc843 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Lock "4b5e5821-9def-4b53-be19-355d9e5f81ec" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.099958] env[63531]: DEBUG oslo_concurrency.lockutils [None req-141fc11b-5904-4c16-af8a-e3b6485bc843 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Acquiring lock "4b5e5821-9def-4b53-be19-355d9e5f81ec-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.100161] env[63531]: DEBUG oslo_concurrency.lockutils [None req-141fc11b-5904-4c16-af8a-e3b6485bc843 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Lock "4b5e5821-9def-4b53-be19-355d9e5f81ec-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.100333] env[63531]: DEBUG oslo_concurrency.lockutils [None req-141fc11b-5904-4c16-af8a-e3b6485bc843 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Lock "4b5e5821-9def-4b53-be19-355d9e5f81ec-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.102559] env[63531]: INFO nova.compute.manager [None req-141fc11b-5904-4c16-af8a-e3b6485bc843 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Terminating instance [ 930.104653] env[63531]: DEBUG nova.compute.manager [None req-141fc11b-5904-4c16-af8a-e3b6485bc843 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 930.104924] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-141fc11b-5904-4c16-af8a-e3b6485bc843 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 930.105720] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d52aa58-6790-4f4b-8b8c-eefc2cf8c444 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.116056] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-141fc11b-5904-4c16-af8a-e3b6485bc843 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 930.116333] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b612ee35-f8ba-449a-a468-a0e4aa501c99 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.125832] env[63531]: DEBUG oslo_vmware.api [None req-141fc11b-5904-4c16-af8a-e3b6485bc843 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Waiting for the task: (returnval){ [ 930.125832] env[63531]: value = "task-1118198" [ 930.125832] env[63531]: _type = "Task" [ 930.125832] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.136175] env[63531]: DEBUG oslo_vmware.api [None req-141fc11b-5904-4c16-af8a-e3b6485bc843 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1118198, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.248185] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-132ff447-284d-4f08-bf28-5afb769e38dd tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 930.248538] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ad3c9c5c-b492-45ba-bf76-74a8c694982a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.259867] env[63531]: DEBUG oslo_vmware.api [None req-132ff447-284d-4f08-bf28-5afb769e38dd tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Waiting for the task: (returnval){ [ 930.259867] env[63531]: value = "task-1118199" [ 930.259867] env[63531]: _type = "Task" [ 930.259867] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.272149] env[63531]: DEBUG oslo_vmware.api [None req-132ff447-284d-4f08-bf28-5afb769e38dd tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118199, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.405243] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 930.405243] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c4c8e2e4-08ba-4bd7-a21b-8f578ae0d817 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.407853] env[63531]: DEBUG nova.compute.utils [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 930.409357] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Releasing lock "refresh_cache-ccf00b09-29fe-4a76-a8af-97774acd77de" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 930.409662] env[63531]: DEBUG nova.compute.manager [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Instance network_info: |[{"id": "7433ce46-ff5a-4e78-bcc9-9dcf2264a01b", "address": "fa:16:3e:ad:77:b5", "network": {"id": "913045d9-e90b-4366-987c-7706b73a6147", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1696510389-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f71c7db0a3614995becc25d9a8dce787", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c297fe21-cd0b-4226-813b-a65d2358d034", "external-id": "nsx-vlan-transportzone-98", "segmentation_id": 98, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7433ce46-ff", "ovs_interfaceid": "7433ce46-ff5a-4e78-bcc9-9dcf2264a01b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 930.410187] env[63531]: DEBUG nova.compute.manager [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3fb0499d-8074-4e63-be9f-380730416cc4] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 930.410372] env[63531]: DEBUG nova.network.neutron [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3fb0499d-8074-4e63-be9f-380730416cc4] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 930.415747] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ad:77:b5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c297fe21-cd0b-4226-813b-a65d2358d034', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7433ce46-ff5a-4e78-bcc9-9dcf2264a01b', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 930.423455] env[63531]: DEBUG oslo.service.loopingcall [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 930.426218] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e904d36d-d4cc-44ab-9e35-5384295ada3e tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 0.523s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.427268] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 930.427958] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.660s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.429641] env[63531]: INFO nova.compute.claims [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 930.435060] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9fb2eaab-7225-404b-b463-18d2f663d752 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.451740] env[63531]: DEBUG oslo_vmware.api [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Waiting for the task: (returnval){ [ 930.451740] env[63531]: value = "task-1118200" [ 930.451740] env[63531]: _type = "Task" [ 930.451740] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.462526] env[63531]: DEBUG oslo_vmware.api [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Task: {'id': task-1118200, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.466811] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 930.466811] env[63531]: value = "task-1118201" [ 930.466811] env[63531]: _type = "Task" [ 930.466811] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.475006] env[63531]: DEBUG oslo_vmware.api [None req-b9944abd-dda8-40ee-b4da-5c9bb59d9201 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118197, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.478520] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118201, 'name': CreateVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.495641] env[63531]: DEBUG nova.policy [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7140e12106504b3e95ff49786361771b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '63900312a30b4942854f552758c039cf', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 930.576691] env[63531]: DEBUG oslo_vmware.api [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': task-1118194, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.638969] env[63531]: DEBUG oslo_vmware.api [None req-141fc11b-5904-4c16-af8a-e3b6485bc843 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1118198, 'name': PowerOffVM_Task, 'duration_secs': 0.206509} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.639373] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-141fc11b-5904-4c16-af8a-e3b6485bc843 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 930.639570] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-141fc11b-5904-4c16-af8a-e3b6485bc843 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 930.639931] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8e37e735-d15d-4ebd-b7b4-3b3158830fd0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.706941] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-141fc11b-5904-4c16-af8a-e3b6485bc843 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 930.707986] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-141fc11b-5904-4c16-af8a-e3b6485bc843 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 930.707986] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-141fc11b-5904-4c16-af8a-e3b6485bc843 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Deleting the datastore file [datastore1] 4b5e5821-9def-4b53-be19-355d9e5f81ec {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 930.707986] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4fc93537-7c61-4190-a630-c22a1c4dbbcf {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.715803] env[63531]: DEBUG oslo_vmware.api [None req-141fc11b-5904-4c16-af8a-e3b6485bc843 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Waiting for the task: (returnval){ [ 930.715803] env[63531]: value = "task-1118203" [ 930.715803] env[63531]: _type = "Task" [ 930.715803] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.725774] env[63531]: DEBUG oslo_vmware.api [None req-141fc11b-5904-4c16-af8a-e3b6485bc843 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1118203, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.771331] env[63531]: DEBUG oslo_vmware.api [None req-132ff447-284d-4f08-bf28-5afb769e38dd tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118199, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.798894] env[63531]: DEBUG nova.network.neutron [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3fb0499d-8074-4e63-be9f-380730416cc4] Successfully created port: 893b680c-06c9-4135-aceb-e275af9a88d7 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 930.911219] env[63531]: DEBUG nova.compute.manager [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3fb0499d-8074-4e63-be9f-380730416cc4] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 930.966212] env[63531]: DEBUG oslo_vmware.api [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Task: {'id': task-1118200, 'name': PowerOffVM_Task} progress is 100%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.978570] env[63531]: DEBUG oslo_vmware.api [None req-b9944abd-dda8-40ee-b4da-5c9bb59d9201 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118197, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.981911] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118201, 'name': CreateVM_Task, 'duration_secs': 0.381452} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.982122] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 930.982832] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 930.982962] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.983327] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 930.983604] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-425d5263-581c-4743-ac27-f0ba88265b85 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.989802] env[63531]: DEBUG oslo_vmware.api [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 930.989802] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52a73e48-662b-6cc5-3722-a174fbcb2096" [ 930.989802] env[63531]: _type = "Task" [ 930.989802] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.999701] env[63531]: DEBUG oslo_vmware.api [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52a73e48-662b-6cc5-3722-a174fbcb2096, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.037932] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ab94488a-b9cf-440f-93f3-139b42979e7b tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "interface-a090135f-489d-457d-be36-ba6f61b71ab8-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.038289] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ab94488a-b9cf-440f-93f3-139b42979e7b tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "interface-a090135f-489d-457d-be36-ba6f61b71ab8-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.038734] env[63531]: DEBUG nova.objects.instance [None req-ab94488a-b9cf-440f-93f3-139b42979e7b tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lazy-loading 'flavor' on Instance uuid a090135f-489d-457d-be36-ba6f61b71ab8 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 931.082814] env[63531]: DEBUG oslo_vmware.api [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': task-1118194, 'name': MoveVirtualDisk_Task} progress is 85%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.227229] env[63531]: DEBUG oslo_vmware.api [None req-141fc11b-5904-4c16-af8a-e3b6485bc843 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Task: {'id': task-1118203, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.21142} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.227522] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-141fc11b-5904-4c16-af8a-e3b6485bc843 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 931.227799] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-141fc11b-5904-4c16-af8a-e3b6485bc843 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 931.227990] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-141fc11b-5904-4c16-af8a-e3b6485bc843 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 931.228191] env[63531]: INFO nova.compute.manager [None req-141fc11b-5904-4c16-af8a-e3b6485bc843 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Took 1.12 seconds to destroy the instance on the hypervisor. [ 931.228441] env[63531]: DEBUG oslo.service.loopingcall [None req-141fc11b-5904-4c16-af8a-e3b6485bc843 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 931.228646] env[63531]: DEBUG nova.compute.manager [-] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 931.228743] env[63531]: DEBUG nova.network.neutron [-] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 931.274284] env[63531]: DEBUG oslo_vmware.api [None req-132ff447-284d-4f08-bf28-5afb769e38dd tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118199, 'name': PowerOnVM_Task, 'duration_secs': 0.794832} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.274604] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-132ff447-284d-4f08-bf28-5afb769e38dd tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 931.274799] env[63531]: DEBUG nova.compute.manager [None req-132ff447-284d-4f08-bf28-5afb769e38dd tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 931.275726] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f7cf209-079d-48aa-a18c-7d1fe296bbc3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.323144] env[63531]: DEBUG nova.network.neutron [None req-ff3c91f0-14af-4f9f-8d99-3fe699fd3d7c tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Updating instance_info_cache with network_info: [{"id": "e4133e0f-55c1-4b3f-847d-899dd1167eec", "address": "fa:16:3e:90:40:9b", "network": {"id": "c4f4c994-36ab-42ab-94f8-78f29cbafa50", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-994507698-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "57e9bf4ee0f1463fa67b0f8deff30fdc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "767a3a48-41d4-4a0c-961d-0024837f63bd", "external-id": "nsx-vlan-transportzone-141", "segmentation_id": 141, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4133e0f-55", "ovs_interfaceid": "e4133e0f-55c1-4b3f-847d-899dd1167eec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 931.463177] env[63531]: DEBUG oslo_vmware.api [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Task: {'id': task-1118200, 'name': PowerOffVM_Task, 'duration_secs': 0.58752} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.463472] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 931.463732] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 931.464500] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f45a47dc-313d-4844-b36e-94f367bd84af {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.479723] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 931.482997] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c88ccef7-b32f-4d23-bfdb-4579ea532c10 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.486059] env[63531]: DEBUG oslo_vmware.api [None req-b9944abd-dda8-40ee-b4da-5c9bb59d9201 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118197, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.498832] env[63531]: DEBUG oslo_vmware.api [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52a73e48-662b-6cc5-3722-a174fbcb2096, 'name': SearchDatastore_Task, 'duration_secs': 0.023834} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.501676] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.501943] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 931.502224] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 931.502482] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.502642] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 931.503820] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-97a6ce26-df84-4bb7-9c63-ad2e1456d070 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.509009] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 931.509237] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Deleting contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 931.509424] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Deleting the datastore file [datastore2] e065f052-ba3e-4783-8953-5dc200d1f3e9 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 931.509668] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2264ef0c-725b-4c4c-84c9-9e9624a15763 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.513264] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 931.513480] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 931.518410] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6e43fa45-0137-4a20-b2f4-c7eca5a84a4b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.518767] env[63531]: DEBUG oslo_vmware.api [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Waiting for the task: (returnval){ [ 931.518767] env[63531]: value = "task-1118205" [ 931.518767] env[63531]: _type = "Task" [ 931.518767] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.525151] env[63531]: DEBUG oslo_vmware.api [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 931.525151] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52574d1d-28b3-abdb-aa4d-b740ffdad929" [ 931.525151] env[63531]: _type = "Task" [ 931.525151] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.531628] env[63531]: DEBUG oslo_vmware.api [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Task: {'id': task-1118205, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.541169] env[63531]: DEBUG oslo_vmware.api [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52574d1d-28b3-abdb-aa4d-b740ffdad929, 'name': SearchDatastore_Task, 'duration_secs': 0.008828} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.544206] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd650bec-c63d-42cd-a4e4-705a408405d1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.551503] env[63531]: DEBUG oslo_vmware.api [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 931.551503] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5260f8ee-b4e0-5124-71e3-9fd702b7f229" [ 931.551503] env[63531]: _type = "Task" [ 931.551503] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.562108] env[63531]: DEBUG oslo_vmware.api [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5260f8ee-b4e0-5124-71e3-9fd702b7f229, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.578407] env[63531]: DEBUG oslo_vmware.api [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': task-1118194, 'name': MoveVirtualDisk_Task, 'duration_secs': 3.000092} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.578822] env[63531]: INFO nova.virt.vmwareapi.ds_util [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_a108e51e-356f-48b7-9db1-3f8ae81f2b55/OSTACK_IMG_a108e51e-356f-48b7-9db1-3f8ae81f2b55.vmdk to [datastore2] devstack-image-cache_base/d477826e-3e39-4ca7-bc24-998780166e1c/d477826e-3e39-4ca7-bc24-998780166e1c.vmdk. [ 931.579040] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Cleaning up location [datastore2] OSTACK_IMG_a108e51e-356f-48b7-9db1-3f8ae81f2b55 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 931.579617] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_a108e51e-356f-48b7-9db1-3f8ae81f2b55 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 931.581267] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a7086386-938f-4d0e-bc58-eaf3f0ee1721 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.590192] env[63531]: DEBUG oslo_vmware.api [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Waiting for the task: (returnval){ [ 931.590192] env[63531]: value = "task-1118206" [ 931.590192] env[63531]: _type = "Task" [ 931.590192] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.600601] env[63531]: DEBUG oslo_vmware.api [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': task-1118206, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.623896] env[63531]: DEBUG nova.objects.instance [None req-ab94488a-b9cf-440f-93f3-139b42979e7b tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lazy-loading 'pci_requests' on Instance uuid a090135f-489d-457d-be36-ba6f61b71ab8 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 931.775562] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-430f2728-78df-447c-b6ac-ed99498a1c10 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.783484] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfd317b6-5078-4a8c-9c00-adb9d6b2ceae {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.818595] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09bd15c1-a51c-449d-9d90-09b329c99ed8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.827364] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c93d059e-97ff-440a-b2e0-d5f56c5ca99c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.831747] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ff3c91f0-14af-4f9f-8d99-3fe699fd3d7c tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Releasing lock "refresh_cache-9a3fad46-a15a-451c-bdab-a3c8cc8add07" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.831979] env[63531]: DEBUG nova.objects.instance [None req-ff3c91f0-14af-4f9f-8d99-3fe699fd3d7c tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lazy-loading 'migration_context' on Instance uuid 9a3fad46-a15a-451c-bdab-a3c8cc8add07 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 931.844955] env[63531]: DEBUG nova.compute.provider_tree [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 931.922574] env[63531]: DEBUG nova.compute.manager [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3fb0499d-8074-4e63-be9f-380730416cc4] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 931.951046] env[63531]: DEBUG nova.virt.hardware [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 931.951046] env[63531]: DEBUG nova.virt.hardware [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 931.951046] env[63531]: DEBUG nova.virt.hardware [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 931.951299] env[63531]: DEBUG nova.virt.hardware [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 931.951345] env[63531]: DEBUG nova.virt.hardware [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 931.951492] env[63531]: DEBUG nova.virt.hardware [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 931.951710] env[63531]: DEBUG nova.virt.hardware [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 931.951873] env[63531]: DEBUG nova.virt.hardware [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 931.952069] env[63531]: DEBUG nova.virt.hardware [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 931.952243] env[63531]: DEBUG nova.virt.hardware [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 931.952426] env[63531]: DEBUG nova.virt.hardware [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 931.953390] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b27f74d7-1fb0-4603-9d64-d01f35020936 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.962418] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97e04986-4961-420c-a1fc-e5bfa2ceab2e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.968541] env[63531]: DEBUG nova.compute.manager [req-e36d8c08-8f22-41b2-b3f3-27169008cbb8 req-93b8e4ad-c2c5-4c1d-bf46-e08abec79df8 service nova] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Received event network-vif-deleted-d31275a8-8fba-48c3-be21-3b90d7cc05cf {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 931.968819] env[63531]: INFO nova.compute.manager [req-e36d8c08-8f22-41b2-b3f3-27169008cbb8 req-93b8e4ad-c2c5-4c1d-bf46-e08abec79df8 service nova] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Neutron deleted interface d31275a8-8fba-48c3-be21-3b90d7cc05cf; detaching it from the instance and deleting it from the info cache [ 931.968969] env[63531]: DEBUG nova.network.neutron [req-e36d8c08-8f22-41b2-b3f3-27169008cbb8 req-93b8e4ad-c2c5-4c1d-bf46-e08abec79df8 service nova] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 931.994050] env[63531]: DEBUG oslo_vmware.api [None req-b9944abd-dda8-40ee-b4da-5c9bb59d9201 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118197, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.546692} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.994417] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9944abd-dda8-40ee-b4da-5c9bb59d9201 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 931.994678] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-b9944abd-dda8-40ee-b4da-5c9bb59d9201 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Deleted contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 931.994933] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-b9944abd-dda8-40ee-b4da-5c9bb59d9201 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 931.995196] env[63531]: INFO nova.compute.manager [None req-b9944abd-dda8-40ee-b4da-5c9bb59d9201 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Took 2.68 seconds to destroy the instance on the hypervisor. [ 931.995538] env[63531]: DEBUG oslo.service.loopingcall [None req-b9944abd-dda8-40ee-b4da-5c9bb59d9201 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 931.995782] env[63531]: DEBUG nova.compute.manager [-] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 931.995910] env[63531]: DEBUG nova.network.neutron [-] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 932.031073] env[63531]: DEBUG oslo_vmware.api [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Task: {'id': task-1118205, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.137325} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.031296] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 932.031489] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Deleted contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 932.031674] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 932.062533] env[63531]: DEBUG oslo_vmware.api [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5260f8ee-b4e0-5124-71e3-9fd702b7f229, 'name': SearchDatastore_Task, 'duration_secs': 0.013047} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.062718] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.063306] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] ccf00b09-29fe-4a76-a8af-97774acd77de/ccf00b09-29fe-4a76-a8af-97774acd77de.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 932.063306] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e163c2c8-6eb2-4f16-aeba-7b223aadbf79 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.070620] env[63531]: DEBUG oslo_vmware.api [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 932.070620] env[63531]: value = "task-1118207" [ 932.070620] env[63531]: _type = "Task" [ 932.070620] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.078911] env[63531]: DEBUG oslo_vmware.api [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118207, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.096332] env[63531]: DEBUG nova.network.neutron [-] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 932.100324] env[63531]: DEBUG oslo_vmware.api [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': task-1118206, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.044483} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.101540] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 932.101540] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d477826e-3e39-4ca7-bc24-998780166e1c/d477826e-3e39-4ca7-bc24-998780166e1c.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.101540] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d477826e-3e39-4ca7-bc24-998780166e1c/d477826e-3e39-4ca7-bc24-998780166e1c.vmdk to [datastore2] 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f/2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 932.101540] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d0d41264-8537-4ce3-b7e9-6993d701a145 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.108379] env[63531]: DEBUG oslo_vmware.api [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Waiting for the task: (returnval){ [ 932.108379] env[63531]: value = "task-1118208" [ 932.108379] env[63531]: _type = "Task" [ 932.108379] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.116615] env[63531]: DEBUG oslo_vmware.api [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': task-1118208, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.129119] env[63531]: DEBUG nova.objects.base [None req-ab94488a-b9cf-440f-93f3-139b42979e7b tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=63531) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 932.129640] env[63531]: DEBUG nova.network.neutron [None req-ab94488a-b9cf-440f-93f3-139b42979e7b tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 932.180305] env[63531]: DEBUG nova.policy [None req-ab94488a-b9cf-440f-93f3-139b42979e7b tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '36cb2f0de54f417384fa1fb2210de4e4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8ba8b284854242c392aec5326e996239', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 932.335287] env[63531]: DEBUG nova.objects.base [None req-ff3c91f0-14af-4f9f-8d99-3fe699fd3d7c tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Object Instance<9a3fad46-a15a-451c-bdab-a3c8cc8add07> lazy-loaded attributes: info_cache,migration_context {{(pid=63531) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 932.337347] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c316661-cc82-49d9-8899-6bbf44364cba {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.361078] env[63531]: DEBUG nova.scheduler.client.report [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 932.366298] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4f07405e-89fc-4b86-bf1d-ca4567a2f484 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.375056] env[63531]: DEBUG oslo_vmware.api [None req-ff3c91f0-14af-4f9f-8d99-3fe699fd3d7c tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 932.375056] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52b6f6f6-ff25-5800-9ac6-ca67abd74735" [ 932.375056] env[63531]: _type = "Task" [ 932.375056] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.387242] env[63531]: DEBUG oslo_vmware.api [None req-ff3c91f0-14af-4f9f-8d99-3fe699fd3d7c tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52b6f6f6-ff25-5800-9ac6-ca67abd74735, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.471804] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-37d723b2-03b1-4ab6-965c-6ad7bbe89166 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.483420] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcdf272c-ee18-44b9-a640-45b37e5d0077 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.497766] env[63531]: DEBUG nova.network.neutron [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3fb0499d-8074-4e63-be9f-380730416cc4] Successfully updated port: 893b680c-06c9-4135-aceb-e275af9a88d7 {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 932.522697] env[63531]: DEBUG nova.compute.manager [req-e36d8c08-8f22-41b2-b3f3-27169008cbb8 req-93b8e4ad-c2c5-4c1d-bf46-e08abec79df8 service nova] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Detach interface failed, port_id=d31275a8-8fba-48c3-be21-3b90d7cc05cf, reason: Instance 4b5e5821-9def-4b53-be19-355d9e5f81ec could not be found. {{(pid=63531) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 932.585316] env[63531]: DEBUG oslo_vmware.api [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118207, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.601257] env[63531]: INFO nova.compute.manager [-] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Took 1.37 seconds to deallocate network for instance. [ 932.621680] env[63531]: DEBUG oslo_vmware.api [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': task-1118208, 'name': CopyVirtualDisk_Task} progress is 18%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.826307] env[63531]: DEBUG nova.network.neutron [-] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 932.868839] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.439s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.868839] env[63531]: DEBUG nova.compute.manager [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 932.870921] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9bd3dc4a-86f3-4a59-85fb-1184d578df53 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.907s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.871182] env[63531]: DEBUG nova.objects.instance [None req-9bd3dc4a-86f3-4a59-85fb-1184d578df53 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lazy-loading 'resources' on Instance uuid dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 932.890924] env[63531]: DEBUG oslo_vmware.api [None req-ff3c91f0-14af-4f9f-8d99-3fe699fd3d7c tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52b6f6f6-ff25-5800-9ac6-ca67abd74735, 'name': SearchDatastore_Task, 'duration_secs': 0.039778} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.891588] env[63531]: DEBUG nova.network.neutron [None req-ab94488a-b9cf-440f-93f3-139b42979e7b tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Successfully created port: 1d5fea6d-10c4-4740-81a1-8e1ac927dad5 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 932.893707] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ff3c91f0-14af-4f9f-8d99-3fe699fd3d7c tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.005841] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "refresh_cache-3fb0499d-8074-4e63-be9f-380730416cc4" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 933.005841] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquired lock "refresh_cache-3fb0499d-8074-4e63-be9f-380730416cc4" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 933.005841] env[63531]: DEBUG nova.network.neutron [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3fb0499d-8074-4e63-be9f-380730416cc4] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 933.036228] env[63531]: DEBUG nova.compute.manager [req-8fd34958-8169-4fa7-ba70-735966126875 req-b41b262b-cd1c-4306-a33e-37a71b918b2c service nova] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Received event network-vif-deleted-0997f575-14bc-4813-9df6-b18e4c1c2186 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 933.079342] env[63531]: DEBUG nova.virt.hardware [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 933.079685] env[63531]: DEBUG nova.virt.hardware [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 933.079896] env[63531]: DEBUG nova.virt.hardware [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 933.080063] env[63531]: DEBUG nova.virt.hardware [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 933.080219] env[63531]: DEBUG nova.virt.hardware [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 933.080371] env[63531]: DEBUG nova.virt.hardware [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 933.080584] env[63531]: DEBUG nova.virt.hardware [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 933.080750] env[63531]: DEBUG nova.virt.hardware [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 933.080925] env[63531]: DEBUG nova.virt.hardware [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 933.081116] env[63531]: DEBUG nova.virt.hardware [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 933.081302] env[63531]: DEBUG nova.virt.hardware [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 933.082533] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47e602b3-1f29-4d23-8277-1c42ab0713b9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.096035] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8042b43-3676-475d-ba96-7c3a4c69d262 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.100147] env[63531]: DEBUG oslo_vmware.api [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118207, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.605295} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.100444] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] ccf00b09-29fe-4a76-a8af-97774acd77de/ccf00b09-29fe-4a76-a8af-97774acd77de.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 933.100651] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 933.101337] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3a447774-1b1c-4465-a84a-b8f32a630fd5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.115721] env[63531]: DEBUG oslo_concurrency.lockutils [None req-141fc11b-5904-4c16-af8a-e3b6485bc843 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.116822] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Instance VIF info [] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 933.122719] env[63531]: DEBUG oslo.service.loopingcall [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 933.127611] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 933.127993] env[63531]: DEBUG oslo_vmware.api [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 933.127993] env[63531]: value = "task-1118209" [ 933.127993] env[63531]: _type = "Task" [ 933.127993] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.128227] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1d3e785d-ed8b-405d-965d-14600ffdd8a7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.149989] env[63531]: DEBUG oslo_vmware.api [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': task-1118208, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.156025] env[63531]: DEBUG oslo_vmware.api [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118209, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.156631] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 933.156631] env[63531]: value = "task-1118210" [ 933.156631] env[63531]: _type = "Task" [ 933.156631] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.167172] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118210, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.328782] env[63531]: INFO nova.compute.manager [-] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Took 1.33 seconds to deallocate network for instance. [ 933.375583] env[63531]: DEBUG nova.compute.utils [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 933.381058] env[63531]: DEBUG nova.compute.manager [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 933.381257] env[63531]: DEBUG nova.network.neutron [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 933.431706] env[63531]: DEBUG nova.policy [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '049cb89a769f41bd91d0e62784635cbd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c2e99349303e413b9f7896d449cb7dff', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 933.561019] env[63531]: DEBUG nova.network.neutron [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3fb0499d-8074-4e63-be9f-380730416cc4] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 933.630599] env[63531]: DEBUG oslo_vmware.api [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': task-1118208, 'name': CopyVirtualDisk_Task} progress is 60%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.653375] env[63531]: DEBUG oslo_vmware.api [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118209, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.37357} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.656414] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 933.657789] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1116772-1053-4261-ba01-60960e594307 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.671466] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118210, 'name': CreateVM_Task, 'duration_secs': 0.393534} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.686271] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 933.695719] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Reconfiguring VM instance instance-00000051 to attach disk [datastore1] ccf00b09-29fe-4a76-a8af-97774acd77de/ccf00b09-29fe-4a76-a8af-97774acd77de.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 933.696640] env[63531]: DEBUG oslo_concurrency.lockutils [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 933.697750] env[63531]: DEBUG oslo_concurrency.lockutils [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 933.697750] env[63531]: DEBUG oslo_concurrency.lockutils [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 933.697750] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3034463a-3570-4bb6-b5a9-970a73983b6e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.715430] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ee68a57f-e57e-41ec-8743-f65171ef8aaf {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.721999] env[63531]: DEBUG oslo_vmware.api [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Waiting for the task: (returnval){ [ 933.721999] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52a5e0a4-8637-311e-027b-a804d25b2802" [ 933.721999] env[63531]: _type = "Task" [ 933.721999] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.727136] env[63531]: DEBUG oslo_vmware.api [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 933.727136] env[63531]: value = "task-1118211" [ 933.727136] env[63531]: _type = "Task" [ 933.727136] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.730577] env[63531]: DEBUG oslo_vmware.api [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52a5e0a4-8637-311e-027b-a804d25b2802, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.742230] env[63531]: DEBUG oslo_vmware.api [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118211, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.762791] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83ee107f-cda7-41a4-9422-a4b4f107c4bd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.770583] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8665ed9-9a10-4ac7-bab5-499d8c6891eb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.805044] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-856b4181-f313-4940-bc43-c0419c3b86cc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.813553] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b99b8562-46f3-4793-9701-9fb90fef3d2a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.827757] env[63531]: DEBUG nova.compute.provider_tree [None req-9bd3dc4a-86f3-4a59-85fb-1184d578df53 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 933.835713] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b9944abd-dda8-40ee-b4da-5c9bb59d9201 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.839792] env[63531]: DEBUG nova.network.neutron [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3fb0499d-8074-4e63-be9f-380730416cc4] Updating instance_info_cache with network_info: [{"id": "893b680c-06c9-4135-aceb-e275af9a88d7", "address": "fa:16:3e:a0:72:3f", "network": {"id": "80641cc0-680b-40ff-8d1b-f5a5f736a88f", "bridge": "br-int", "label": "tempest-ServersTestJSON-666288463-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "63900312a30b4942854f552758c039cf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "26472e27-9835-4f87-ab7f-ca24dfee4e83", "external-id": "nsx-vlan-transportzone-335", "segmentation_id": 335, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap893b680c-06", "ovs_interfaceid": "893b680c-06c9-4135-aceb-e275af9a88d7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 933.851401] env[63531]: DEBUG nova.network.neutron [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Successfully created port: a755a88f-c86c-4183-b9fa-a6133060ea5e {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 933.881933] env[63531]: DEBUG nova.compute.manager [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 934.127646] env[63531]: DEBUG oslo_vmware.api [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': task-1118208, 'name': CopyVirtualDisk_Task} progress is 80%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.232110] env[63531]: DEBUG oslo_vmware.api [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52a5e0a4-8637-311e-027b-a804d25b2802, 'name': SearchDatastore_Task, 'duration_secs': 0.013311} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.235393] env[63531]: DEBUG oslo_concurrency.lockutils [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.235665] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 934.235916] env[63531]: DEBUG oslo_concurrency.lockutils [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.236088] env[63531]: DEBUG oslo_concurrency.lockutils [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.236283] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 934.236566] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-74222f42-2b1c-4ba1-9564-3887210a9b0d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.243835] env[63531]: DEBUG oslo_vmware.api [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118211, 'name': ReconfigVM_Task, 'duration_secs': 0.332881} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.244127] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Reconfigured VM instance instance-00000051 to attach disk [datastore1] ccf00b09-29fe-4a76-a8af-97774acd77de/ccf00b09-29fe-4a76-a8af-97774acd77de.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 934.244767] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-43908a4a-0716-42a3-adc4-fd6d06d27745 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.247414] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 934.247627] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 934.248720] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6d8c6f36-33c6-42a2-afed-fe2af32a26d6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.253257] env[63531]: DEBUG oslo_vmware.api [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 934.253257] env[63531]: value = "task-1118212" [ 934.253257] env[63531]: _type = "Task" [ 934.253257] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.257757] env[63531]: DEBUG oslo_vmware.api [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Waiting for the task: (returnval){ [ 934.257757] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]524c70e8-70c0-a527-4d04-b4098e00c2c6" [ 934.257757] env[63531]: _type = "Task" [ 934.257757] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.264287] env[63531]: DEBUG oslo_vmware.api [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118212, 'name': Rename_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.271581] env[63531]: DEBUG oslo_vmware.api [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]524c70e8-70c0-a527-4d04-b4098e00c2c6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.331572] env[63531]: DEBUG nova.scheduler.client.report [None req-9bd3dc4a-86f3-4a59-85fb-1184d578df53 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 934.342058] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Releasing lock "refresh_cache-3fb0499d-8074-4e63-be9f-380730416cc4" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.342058] env[63531]: DEBUG nova.compute.manager [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3fb0499d-8074-4e63-be9f-380730416cc4] Instance network_info: |[{"id": "893b680c-06c9-4135-aceb-e275af9a88d7", "address": "fa:16:3e:a0:72:3f", "network": {"id": "80641cc0-680b-40ff-8d1b-f5a5f736a88f", "bridge": "br-int", "label": "tempest-ServersTestJSON-666288463-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "63900312a30b4942854f552758c039cf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "26472e27-9835-4f87-ab7f-ca24dfee4e83", "external-id": "nsx-vlan-transportzone-335", "segmentation_id": 335, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap893b680c-06", "ovs_interfaceid": "893b680c-06c9-4135-aceb-e275af9a88d7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 934.342393] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3fb0499d-8074-4e63-be9f-380730416cc4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a0:72:3f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '26472e27-9835-4f87-ab7f-ca24dfee4e83', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '893b680c-06c9-4135-aceb-e275af9a88d7', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 934.350436] env[63531]: DEBUG oslo.service.loopingcall [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 934.351494] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3fb0499d-8074-4e63-be9f-380730416cc4] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 934.351811] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-aed4b151-5071-4e90-8098-149ed00488d2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.375962] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 934.375962] env[63531]: value = "task-1118213" [ 934.375962] env[63531]: _type = "Task" [ 934.375962] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.389943] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118213, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.625854] env[63531]: DEBUG oslo_vmware.api [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': task-1118208, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.476422} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.626192] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d477826e-3e39-4ca7-bc24-998780166e1c/d477826e-3e39-4ca7-bc24-998780166e1c.vmdk to [datastore2] 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f/2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 934.627029] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fae6ed1-dcd3-4a83-bced-9f16d3f3fdff {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.649850] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Reconfiguring VM instance instance-00000039 to attach disk [datastore2] 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f/2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f.vmdk or device None with type streamOptimized {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 934.650233] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-902ecf10-e5ef-4489-b330-425524d581b6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.668568] env[63531]: DEBUG nova.network.neutron [None req-ab94488a-b9cf-440f-93f3-139b42979e7b tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Successfully updated port: 1d5fea6d-10c4-4740-81a1-8e1ac927dad5 {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 934.671385] env[63531]: DEBUG oslo_vmware.api [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Waiting for the task: (returnval){ [ 934.671385] env[63531]: value = "task-1118214" [ 934.671385] env[63531]: _type = "Task" [ 934.671385] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.679977] env[63531]: DEBUG oslo_vmware.api [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': task-1118214, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.764596] env[63531]: DEBUG oslo_vmware.api [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118212, 'name': Rename_Task, 'duration_secs': 0.154973} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.765248] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 934.765513] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-933586ae-d59f-4b4e-847e-5ebdbdac023f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.770062] env[63531]: DEBUG oslo_vmware.api [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]524c70e8-70c0-a527-4d04-b4098e00c2c6, 'name': SearchDatastore_Task, 'duration_secs': 0.021722} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.771084] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-539a2557-a794-4b35-88ea-27ebc2ecf55e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.774136] env[63531]: DEBUG oslo_vmware.api [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 934.774136] env[63531]: value = "task-1118215" [ 934.774136] env[63531]: _type = "Task" [ 934.774136] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.780227] env[63531]: DEBUG oslo_vmware.api [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Waiting for the task: (returnval){ [ 934.780227] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52683b0d-ff04-c900-6d3e-f21f658147d3" [ 934.780227] env[63531]: _type = "Task" [ 934.780227] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.786033] env[63531]: DEBUG oslo_vmware.api [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118215, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.790903] env[63531]: DEBUG oslo_vmware.api [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52683b0d-ff04-c900-6d3e-f21f658147d3, 'name': SearchDatastore_Task, 'duration_secs': 0.008352} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.791158] env[63531]: DEBUG oslo_concurrency.lockutils [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.791414] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] e065f052-ba3e-4783-8953-5dc200d1f3e9/e065f052-ba3e-4783-8953-5dc200d1f3e9.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 934.791696] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8b7f30d0-b12f-45cf-9e91-8e3a8054fca6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.798224] env[63531]: DEBUG oslo_vmware.api [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Waiting for the task: (returnval){ [ 934.798224] env[63531]: value = "task-1118216" [ 934.798224] env[63531]: _type = "Task" [ 934.798224] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.807019] env[63531]: DEBUG oslo_vmware.api [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Task: {'id': task-1118216, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.836609] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9bd3dc4a-86f3-4a59-85fb-1184d578df53 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.966s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.839110] env[63531]: DEBUG oslo_concurrency.lockutils [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.293s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.840674] env[63531]: INFO nova.compute.claims [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] [instance: 55111b87-1e9a-4877-8689-987faa72a54e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 934.867929] env[63531]: INFO nova.scheduler.client.report [None req-9bd3dc4a-86f3-4a59-85fb-1184d578df53 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Deleted allocations for instance dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a [ 934.885644] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118213, 'name': CreateVM_Task, 'duration_secs': 0.389805} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.885825] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3fb0499d-8074-4e63-be9f-380730416cc4] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 934.886574] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.886765] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.887114] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 934.887374] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4130fa86-8df1-48d2-869a-a48b66860217 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.892570] env[63531]: DEBUG nova.compute.manager [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 934.896203] env[63531]: DEBUG oslo_vmware.api [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 934.896203] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52814115-60ef-bd16-480f-b50d441f9ec8" [ 934.896203] env[63531]: _type = "Task" [ 934.896203] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.906301] env[63531]: DEBUG oslo_vmware.api [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52814115-60ef-bd16-480f-b50d441f9ec8, 'name': SearchDatastore_Task, 'duration_secs': 0.009108} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.906758] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.906758] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3fb0499d-8074-4e63-be9f-380730416cc4] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 934.907211] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.907211] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.907361] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 934.907678] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-39a670cc-308c-4229-9efb-1f98dcd10e76 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.916549] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 934.916759] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 934.917606] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-de84d46f-c903-46a3-b16e-941486c65895 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.922533] env[63531]: DEBUG nova.virt.hardware [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 934.922797] env[63531]: DEBUG nova.virt.hardware [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 934.922968] env[63531]: DEBUG nova.virt.hardware [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 934.923195] env[63531]: DEBUG nova.virt.hardware [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 934.923388] env[63531]: DEBUG nova.virt.hardware [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 934.923555] env[63531]: DEBUG nova.virt.hardware [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 934.923806] env[63531]: DEBUG nova.virt.hardware [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 934.924016] env[63531]: DEBUG nova.virt.hardware [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 934.924216] env[63531]: DEBUG nova.virt.hardware [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 934.924399] env[63531]: DEBUG nova.virt.hardware [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 934.924621] env[63531]: DEBUG nova.virt.hardware [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 934.925690] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-958da981-80c1-4cb0-a014-48458da42878 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.934955] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e74bd9f-1444-4d27-97ad-d552e384484f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.939849] env[63531]: DEBUG oslo_vmware.api [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 934.939849] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]524ebb41-2e2b-cf1d-395f-47d20dc2db26" [ 934.939849] env[63531]: _type = "Task" [ 934.939849] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.956750] env[63531]: DEBUG nova.compute.manager [req-47d690a6-5c2e-475e-b6c7-2b8562558239 req-5b7ef332-486a-4bd6-a837-7a8cf99c382a service nova] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Received event network-vif-plugged-1d5fea6d-10c4-4740-81a1-8e1ac927dad5 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 934.957051] env[63531]: DEBUG oslo_concurrency.lockutils [req-47d690a6-5c2e-475e-b6c7-2b8562558239 req-5b7ef332-486a-4bd6-a837-7a8cf99c382a service nova] Acquiring lock "a090135f-489d-457d-be36-ba6f61b71ab8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.957299] env[63531]: DEBUG oslo_concurrency.lockutils [req-47d690a6-5c2e-475e-b6c7-2b8562558239 req-5b7ef332-486a-4bd6-a837-7a8cf99c382a service nova] Lock "a090135f-489d-457d-be36-ba6f61b71ab8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.957498] env[63531]: DEBUG oslo_concurrency.lockutils [req-47d690a6-5c2e-475e-b6c7-2b8562558239 req-5b7ef332-486a-4bd6-a837-7a8cf99c382a service nova] Lock "a090135f-489d-457d-be36-ba6f61b71ab8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.957724] env[63531]: DEBUG nova.compute.manager [req-47d690a6-5c2e-475e-b6c7-2b8562558239 req-5b7ef332-486a-4bd6-a837-7a8cf99c382a service nova] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] No waiting events found dispatching network-vif-plugged-1d5fea6d-10c4-4740-81a1-8e1ac927dad5 {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 934.957947] env[63531]: WARNING nova.compute.manager [req-47d690a6-5c2e-475e-b6c7-2b8562558239 req-5b7ef332-486a-4bd6-a837-7a8cf99c382a service nova] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Received unexpected event network-vif-plugged-1d5fea6d-10c4-4740-81a1-8e1ac927dad5 for instance with vm_state active and task_state None. [ 934.962610] env[63531]: DEBUG oslo_vmware.api [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]524ebb41-2e2b-cf1d-395f-47d20dc2db26, 'name': SearchDatastore_Task, 'duration_secs': 0.00971} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.963490] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b079004c-25cb-45b1-859d-885304ad0a59 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.969572] env[63531]: DEBUG oslo_vmware.api [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 934.969572] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]528717cd-7590-0f65-27cb-4d14071ff0f0" [ 934.969572] env[63531]: _type = "Task" [ 934.969572] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.979231] env[63531]: DEBUG oslo_vmware.api [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]528717cd-7590-0f65-27cb-4d14071ff0f0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.171097] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ab94488a-b9cf-440f-93f3-139b42979e7b tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "refresh_cache-a090135f-489d-457d-be36-ba6f61b71ab8" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.171462] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ab94488a-b9cf-440f-93f3-139b42979e7b tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquired lock "refresh_cache-a090135f-489d-457d-be36-ba6f61b71ab8" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.171701] env[63531]: DEBUG nova.network.neutron [None req-ab94488a-b9cf-440f-93f3-139b42979e7b tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 935.183476] env[63531]: DEBUG oslo_vmware.api [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': task-1118214, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.285040] env[63531]: DEBUG oslo_vmware.api [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118215, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.309865] env[63531]: DEBUG oslo_vmware.api [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Task: {'id': task-1118216, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.481452} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.310189] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] e065f052-ba3e-4783-8953-5dc200d1f3e9/e065f052-ba3e-4783-8953-5dc200d1f3e9.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 935.310406] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 935.310660] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-03c236c1-f174-4fb9-893e-28505011c909 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.317410] env[63531]: DEBUG oslo_vmware.api [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Waiting for the task: (returnval){ [ 935.317410] env[63531]: value = "task-1118217" [ 935.317410] env[63531]: _type = "Task" [ 935.317410] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.325588] env[63531]: DEBUG oslo_vmware.api [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Task: {'id': task-1118217, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.378846] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9bd3dc4a-86f3-4a59-85fb-1184d578df53 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.450s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.449499] env[63531]: DEBUG nova.compute.manager [req-390a7a62-dc1d-4f1f-9c7c-ffae55cc4a63 req-8188d500-f504-4926-ae01-2acb267595ff service nova] [instance: 3fb0499d-8074-4e63-be9f-380730416cc4] Received event network-vif-plugged-893b680c-06c9-4135-aceb-e275af9a88d7 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 935.449499] env[63531]: DEBUG oslo_concurrency.lockutils [req-390a7a62-dc1d-4f1f-9c7c-ffae55cc4a63 req-8188d500-f504-4926-ae01-2acb267595ff service nova] Acquiring lock "3fb0499d-8074-4e63-be9f-380730416cc4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.449499] env[63531]: DEBUG oslo_concurrency.lockutils [req-390a7a62-dc1d-4f1f-9c7c-ffae55cc4a63 req-8188d500-f504-4926-ae01-2acb267595ff service nova] Lock "3fb0499d-8074-4e63-be9f-380730416cc4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.449499] env[63531]: DEBUG oslo_concurrency.lockutils [req-390a7a62-dc1d-4f1f-9c7c-ffae55cc4a63 req-8188d500-f504-4926-ae01-2acb267595ff service nova] Lock "3fb0499d-8074-4e63-be9f-380730416cc4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.449944] env[63531]: DEBUG nova.compute.manager [req-390a7a62-dc1d-4f1f-9c7c-ffae55cc4a63 req-8188d500-f504-4926-ae01-2acb267595ff service nova] [instance: 3fb0499d-8074-4e63-be9f-380730416cc4] No waiting events found dispatching network-vif-plugged-893b680c-06c9-4135-aceb-e275af9a88d7 {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 935.450048] env[63531]: WARNING nova.compute.manager [req-390a7a62-dc1d-4f1f-9c7c-ffae55cc4a63 req-8188d500-f504-4926-ae01-2acb267595ff service nova] [instance: 3fb0499d-8074-4e63-be9f-380730416cc4] Received unexpected event network-vif-plugged-893b680c-06c9-4135-aceb-e275af9a88d7 for instance with vm_state building and task_state spawning. [ 935.450297] env[63531]: DEBUG nova.compute.manager [req-390a7a62-dc1d-4f1f-9c7c-ffae55cc4a63 req-8188d500-f504-4926-ae01-2acb267595ff service nova] [instance: 3fb0499d-8074-4e63-be9f-380730416cc4] Received event network-changed-893b680c-06c9-4135-aceb-e275af9a88d7 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 935.450468] env[63531]: DEBUG nova.compute.manager [req-390a7a62-dc1d-4f1f-9c7c-ffae55cc4a63 req-8188d500-f504-4926-ae01-2acb267595ff service nova] [instance: 3fb0499d-8074-4e63-be9f-380730416cc4] Refreshing instance network info cache due to event network-changed-893b680c-06c9-4135-aceb-e275af9a88d7. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 935.450713] env[63531]: DEBUG oslo_concurrency.lockutils [req-390a7a62-dc1d-4f1f-9c7c-ffae55cc4a63 req-8188d500-f504-4926-ae01-2acb267595ff service nova] Acquiring lock "refresh_cache-3fb0499d-8074-4e63-be9f-380730416cc4" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.450904] env[63531]: DEBUG oslo_concurrency.lockutils [req-390a7a62-dc1d-4f1f-9c7c-ffae55cc4a63 req-8188d500-f504-4926-ae01-2acb267595ff service nova] Acquired lock "refresh_cache-3fb0499d-8074-4e63-be9f-380730416cc4" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.451110] env[63531]: DEBUG nova.network.neutron [req-390a7a62-dc1d-4f1f-9c7c-ffae55cc4a63 req-8188d500-f504-4926-ae01-2acb267595ff service nova] [instance: 3fb0499d-8074-4e63-be9f-380730416cc4] Refreshing network info cache for port 893b680c-06c9-4135-aceb-e275af9a88d7 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 935.459086] env[63531]: DEBUG nova.network.neutron [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Successfully updated port: a755a88f-c86c-4183-b9fa-a6133060ea5e {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 935.481900] env[63531]: DEBUG oslo_vmware.api [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]528717cd-7590-0f65-27cb-4d14071ff0f0, 'name': SearchDatastore_Task, 'duration_secs': 0.067638} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.482182] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.482488] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 3fb0499d-8074-4e63-be9f-380730416cc4/3fb0499d-8074-4e63-be9f-380730416cc4.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 935.482946] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-421cf359-625b-4db4-a623-b2249f723519 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.489642] env[63531]: DEBUG oslo_vmware.api [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 935.489642] env[63531]: value = "task-1118218" [ 935.489642] env[63531]: _type = "Task" [ 935.489642] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.497869] env[63531]: DEBUG oslo_vmware.api [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118218, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.683164] env[63531]: DEBUG oslo_vmware.api [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': task-1118214, 'name': ReconfigVM_Task, 'duration_secs': 0.53852} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.684040] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Reconfigured VM instance instance-00000039 to attach disk [datastore2] 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f/2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f.vmdk or device None with type streamOptimized {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 935.684705] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-75821dc7-7147-49d4-b749-4c4dd89f0b48 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.692499] env[63531]: DEBUG oslo_vmware.api [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Waiting for the task: (returnval){ [ 935.692499] env[63531]: value = "task-1118219" [ 935.692499] env[63531]: _type = "Task" [ 935.692499] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.701374] env[63531]: DEBUG oslo_vmware.api [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': task-1118219, 'name': Rename_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.724993] env[63531]: WARNING nova.network.neutron [None req-ab94488a-b9cf-440f-93f3-139b42979e7b tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] caf86428-59a5-4e50-9135-2ab8473af64d already exists in list: networks containing: ['caf86428-59a5-4e50-9135-2ab8473af64d']. ignoring it [ 935.785561] env[63531]: DEBUG oslo_vmware.api [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118215, 'name': PowerOnVM_Task, 'duration_secs': 0.521312} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.788469] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 935.788729] env[63531]: INFO nova.compute.manager [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Took 8.73 seconds to spawn the instance on the hypervisor. [ 935.788930] env[63531]: DEBUG nova.compute.manager [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 935.789851] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73653dba-224b-4b07-b681-5d4c3e9a64dd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.830194] env[63531]: DEBUG oslo_vmware.api [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Task: {'id': task-1118217, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.264016} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.830505] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 935.831389] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09c1fee4-e664-48df-876e-799fe92f47c2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.852350] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Reconfiguring VM instance instance-00000050 to attach disk [datastore1] e065f052-ba3e-4783-8953-5dc200d1f3e9/e065f052-ba3e-4783-8953-5dc200d1f3e9.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 935.856483] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-32971dff-24ac-465b-b37b-452cc73145db {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.880562] env[63531]: DEBUG oslo_vmware.api [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Waiting for the task: (returnval){ [ 935.880562] env[63531]: value = "task-1118220" [ 935.880562] env[63531]: _type = "Task" [ 935.880562] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.892405] env[63531]: DEBUG oslo_vmware.api [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Task: {'id': task-1118220, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.963821] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "refresh_cache-27c23b7c-a68f-43c9-a125-bd657feb3c5b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.964075] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquired lock "refresh_cache-27c23b7c-a68f-43c9-a125-bd657feb3c5b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.964190] env[63531]: DEBUG nova.network.neutron [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 935.999808] env[63531]: DEBUG oslo_vmware.api [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118218, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.431573} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.002127] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 3fb0499d-8074-4e63-be9f-380730416cc4/3fb0499d-8074-4e63-be9f-380730416cc4.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 936.002390] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3fb0499d-8074-4e63-be9f-380730416cc4] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 936.004717] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-198b3fa8-3978-43e9-b6d8-5540d6ef9d30 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.013839] env[63531]: DEBUG oslo_vmware.api [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 936.013839] env[63531]: value = "task-1118221" [ 936.013839] env[63531]: _type = "Task" [ 936.013839] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.023281] env[63531]: DEBUG oslo_vmware.api [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118221, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.088548] env[63531]: DEBUG nova.network.neutron [None req-ab94488a-b9cf-440f-93f3-139b42979e7b tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Updating instance_info_cache with network_info: [{"id": "17da826c-920c-4044-93cb-be8c8cf01c13", "address": "fa:16:3e:9a:ac:ad", "network": {"id": "caf86428-59a5-4e50-9135-2ab8473af64d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1673994696-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.214", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ba8b284854242c392aec5326e996239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap17da826c-92", "ovs_interfaceid": "17da826c-920c-4044-93cb-be8c8cf01c13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "1d5fea6d-10c4-4740-81a1-8e1ac927dad5", "address": "fa:16:3e:06:1e:fa", "network": {"id": "caf86428-59a5-4e50-9135-2ab8473af64d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1673994696-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ba8b284854242c392aec5326e996239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d5fea6d-10", "ovs_interfaceid": "1d5fea6d-10c4-4740-81a1-8e1ac927dad5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.179595] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61108958-943d-4127-9cef-7d3f04b2ec36 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.187617] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fda3259-5e16-414c-8e7a-db7559371f0b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.221445] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac5fb3e6-8598-4d16-90ec-d1c451456a40 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.226849] env[63531]: DEBUG oslo_vmware.api [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': task-1118219, 'name': Rename_Task, 'duration_secs': 0.27892} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.227475] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 936.227774] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-522f22d8-882a-4b8a-bca8-62ed4fad29db {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.232451] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77a53d9b-8f65-4d6b-80f9-4dcfd2c7ed99 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.238619] env[63531]: DEBUG oslo_vmware.api [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Waiting for the task: (returnval){ [ 936.238619] env[63531]: value = "task-1118222" [ 936.238619] env[63531]: _type = "Task" [ 936.238619] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.251333] env[63531]: DEBUG nova.compute.provider_tree [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 936.257793] env[63531]: DEBUG oslo_vmware.api [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': task-1118222, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.275921] env[63531]: DEBUG nova.network.neutron [req-390a7a62-dc1d-4f1f-9c7c-ffae55cc4a63 req-8188d500-f504-4926-ae01-2acb267595ff service nova] [instance: 3fb0499d-8074-4e63-be9f-380730416cc4] Updated VIF entry in instance network info cache for port 893b680c-06c9-4135-aceb-e275af9a88d7. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 936.276371] env[63531]: DEBUG nova.network.neutron [req-390a7a62-dc1d-4f1f-9c7c-ffae55cc4a63 req-8188d500-f504-4926-ae01-2acb267595ff service nova] [instance: 3fb0499d-8074-4e63-be9f-380730416cc4] Updating instance_info_cache with network_info: [{"id": "893b680c-06c9-4135-aceb-e275af9a88d7", "address": "fa:16:3e:a0:72:3f", "network": {"id": "80641cc0-680b-40ff-8d1b-f5a5f736a88f", "bridge": "br-int", "label": "tempest-ServersTestJSON-666288463-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "63900312a30b4942854f552758c039cf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "26472e27-9835-4f87-ab7f-ca24dfee4e83", "external-id": "nsx-vlan-transportzone-335", "segmentation_id": 335, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap893b680c-06", "ovs_interfaceid": "893b680c-06c9-4135-aceb-e275af9a88d7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.308924] env[63531]: INFO nova.compute.manager [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Took 35.45 seconds to build instance. [ 936.391436] env[63531]: DEBUG oslo_vmware.api [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Task: {'id': task-1118220, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.492914] env[63531]: DEBUG nova.network.neutron [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 936.523038] env[63531]: DEBUG oslo_vmware.api [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118221, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065317} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.523363] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3fb0499d-8074-4e63-be9f-380730416cc4] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 936.524233] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78c114c4-c2d0-454e-817d-5bacc5add27a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.545537] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3fb0499d-8074-4e63-be9f-380730416cc4] Reconfiguring VM instance instance-00000052 to attach disk [datastore1] 3fb0499d-8074-4e63-be9f-380730416cc4/3fb0499d-8074-4e63-be9f-380730416cc4.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 936.545823] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ec4159f9-e04e-4e70-9caf-9dd36ec90555 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.567209] env[63531]: DEBUG oslo_vmware.api [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 936.567209] env[63531]: value = "task-1118223" [ 936.567209] env[63531]: _type = "Task" [ 936.567209] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.574653] env[63531]: DEBUG oslo_vmware.api [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118223, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.590918] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ab94488a-b9cf-440f-93f3-139b42979e7b tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Releasing lock "refresh_cache-a090135f-489d-457d-be36-ba6f61b71ab8" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.591589] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ab94488a-b9cf-440f-93f3-139b42979e7b tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "a090135f-489d-457d-be36-ba6f61b71ab8" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.591759] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ab94488a-b9cf-440f-93f3-139b42979e7b tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquired lock "a090135f-489d-457d-be36-ba6f61b71ab8" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.592621] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9d96309-a18a-429d-9161-220b697af97a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.609259] env[63531]: DEBUG nova.virt.hardware [None req-ab94488a-b9cf-440f-93f3-139b42979e7b tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 936.609526] env[63531]: DEBUG nova.virt.hardware [None req-ab94488a-b9cf-440f-93f3-139b42979e7b tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 936.609683] env[63531]: DEBUG nova.virt.hardware [None req-ab94488a-b9cf-440f-93f3-139b42979e7b tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 936.609881] env[63531]: DEBUG nova.virt.hardware [None req-ab94488a-b9cf-440f-93f3-139b42979e7b tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 936.610032] env[63531]: DEBUG nova.virt.hardware [None req-ab94488a-b9cf-440f-93f3-139b42979e7b tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 936.610218] env[63531]: DEBUG nova.virt.hardware [None req-ab94488a-b9cf-440f-93f3-139b42979e7b tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 936.610523] env[63531]: DEBUG nova.virt.hardware [None req-ab94488a-b9cf-440f-93f3-139b42979e7b tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 936.610736] env[63531]: DEBUG nova.virt.hardware [None req-ab94488a-b9cf-440f-93f3-139b42979e7b tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 936.610929] env[63531]: DEBUG nova.virt.hardware [None req-ab94488a-b9cf-440f-93f3-139b42979e7b tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 936.611117] env[63531]: DEBUG nova.virt.hardware [None req-ab94488a-b9cf-440f-93f3-139b42979e7b tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 936.611312] env[63531]: DEBUG nova.virt.hardware [None req-ab94488a-b9cf-440f-93f3-139b42979e7b tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 936.617511] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-ab94488a-b9cf-440f-93f3-139b42979e7b tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Reconfiguring VM to attach interface {{(pid=63531) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 936.620398] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c09b6fea-ef2d-4537-9ff0-333a7175fd1e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.637743] env[63531]: DEBUG oslo_vmware.api [None req-ab94488a-b9cf-440f-93f3-139b42979e7b tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for the task: (returnval){ [ 936.637743] env[63531]: value = "task-1118224" [ 936.637743] env[63531]: _type = "Task" [ 936.637743] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.645764] env[63531]: DEBUG oslo_vmware.api [None req-ab94488a-b9cf-440f-93f3-139b42979e7b tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118224, 'name': ReconfigVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.737014] env[63531]: DEBUG nova.network.neutron [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Updating instance_info_cache with network_info: [{"id": "a755a88f-c86c-4183-b9fa-a6133060ea5e", "address": "fa:16:3e:26:e6:a5", "network": {"id": "bf99b8cb-9efb-4a74-9db1-0f5f0875f26a", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-998205595-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2e99349303e413b9f7896d449cb7dff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc6e6fe1-c4f5-4389-a49f-0978060eebb4", "external-id": "nsx-vlan-transportzone-829", "segmentation_id": 829, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa755a88f-c8", "ovs_interfaceid": "a755a88f-c86c-4183-b9fa-a6133060ea5e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.748766] env[63531]: DEBUG oslo_vmware.api [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': task-1118222, 'name': PowerOnVM_Task, 'duration_secs': 0.470958} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.749713] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 936.754042] env[63531]: DEBUG nova.scheduler.client.report [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 936.779290] env[63531]: DEBUG oslo_concurrency.lockutils [req-390a7a62-dc1d-4f1f-9c7c-ffae55cc4a63 req-8188d500-f504-4926-ae01-2acb267595ff service nova] Releasing lock "refresh_cache-3fb0499d-8074-4e63-be9f-380730416cc4" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.811448] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c82d4def-56f1-4ec5-a600-c6a3c9187383 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Lock "ccf00b09-29fe-4a76-a8af-97774acd77de" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.967s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.850765] env[63531]: DEBUG nova.compute.manager [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 936.851818] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a23482e2-bbda-4124-a67c-7f5f214742e6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.891062] env[63531]: DEBUG oslo_vmware.api [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Task: {'id': task-1118220, 'name': ReconfigVM_Task, 'duration_secs': 0.937341} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.891379] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Reconfigured VM instance instance-00000050 to attach disk [datastore1] e065f052-ba3e-4783-8953-5dc200d1f3e9/e065f052-ba3e-4783-8953-5dc200d1f3e9.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 936.891994] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-37149c84-3004-4615-89ec-4ed63ae6a55b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.898264] env[63531]: DEBUG oslo_vmware.api [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Waiting for the task: (returnval){ [ 936.898264] env[63531]: value = "task-1118225" [ 936.898264] env[63531]: _type = "Task" [ 936.898264] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.908582] env[63531]: DEBUG oslo_vmware.api [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Task: {'id': task-1118225, 'name': Rename_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.078857] env[63531]: DEBUG oslo_vmware.api [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118223, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.147745] env[63531]: DEBUG oslo_vmware.api [None req-ab94488a-b9cf-440f-93f3-139b42979e7b tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118224, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.220659] env[63531]: DEBUG nova.compute.manager [req-e5b432cd-0fea-4d4e-91e1-e023fcbca757 req-8249d2de-c9c0-42ce-b0d7-0f6bf498d028 service nova] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Received event network-changed-1d5fea6d-10c4-4740-81a1-8e1ac927dad5 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 937.220975] env[63531]: DEBUG nova.compute.manager [req-e5b432cd-0fea-4d4e-91e1-e023fcbca757 req-8249d2de-c9c0-42ce-b0d7-0f6bf498d028 service nova] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Refreshing instance network info cache due to event network-changed-1d5fea6d-10c4-4740-81a1-8e1ac927dad5. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 937.221302] env[63531]: DEBUG oslo_concurrency.lockutils [req-e5b432cd-0fea-4d4e-91e1-e023fcbca757 req-8249d2de-c9c0-42ce-b0d7-0f6bf498d028 service nova] Acquiring lock "refresh_cache-a090135f-489d-457d-be36-ba6f61b71ab8" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.221526] env[63531]: DEBUG oslo_concurrency.lockutils [req-e5b432cd-0fea-4d4e-91e1-e023fcbca757 req-8249d2de-c9c0-42ce-b0d7-0f6bf498d028 service nova] Acquired lock "refresh_cache-a090135f-489d-457d-be36-ba6f61b71ab8" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.221740] env[63531]: DEBUG nova.network.neutron [req-e5b432cd-0fea-4d4e-91e1-e023fcbca757 req-8249d2de-c9c0-42ce-b0d7-0f6bf498d028 service nova] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Refreshing network info cache for port 1d5fea6d-10c4-4740-81a1-8e1ac927dad5 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 937.239909] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Releasing lock "refresh_cache-27c23b7c-a68f-43c9-a125-bd657feb3c5b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.240227] env[63531]: DEBUG nova.compute.manager [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Instance network_info: |[{"id": "a755a88f-c86c-4183-b9fa-a6133060ea5e", "address": "fa:16:3e:26:e6:a5", "network": {"id": "bf99b8cb-9efb-4a74-9db1-0f5f0875f26a", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-998205595-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2e99349303e413b9f7896d449cb7dff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc6e6fe1-c4f5-4389-a49f-0978060eebb4", "external-id": "nsx-vlan-transportzone-829", "segmentation_id": 829, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa755a88f-c8", "ovs_interfaceid": "a755a88f-c86c-4183-b9fa-a6133060ea5e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 937.240657] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:26:e6:a5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bc6e6fe1-c4f5-4389-a49f-0978060eebb4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a755a88f-c86c-4183-b9fa-a6133060ea5e', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 937.248428] env[63531]: DEBUG oslo.service.loopingcall [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 937.248999] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 937.249308] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-63eef77b-2378-4b1e-9669-0eb98ddb45cb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.264650] env[63531]: DEBUG oslo_concurrency.lockutils [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.426s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.265198] env[63531]: DEBUG nova.compute.manager [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] [instance: 55111b87-1e9a-4877-8689-987faa72a54e] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 937.267862] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4fccfddd-dc56-46bb-aba7-1e6242d29900 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.471s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.268105] env[63531]: DEBUG nova.objects.instance [None req-4fccfddd-dc56-46bb-aba7-1e6242d29900 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lazy-loading 'resources' on Instance uuid 06ec91b2-14cb-4db8-8fab-e6c4183b1116 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 937.274735] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 937.274735] env[63531]: value = "task-1118226" [ 937.274735] env[63531]: _type = "Task" [ 937.274735] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.283849] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118226, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.372690] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3ca43549-c7ca-4ca5-a29a-7a0f834a0a91 tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Lock "2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 40.567s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.408669] env[63531]: DEBUG oslo_vmware.api [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Task: {'id': task-1118225, 'name': Rename_Task, 'duration_secs': 0.500033} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.408987] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 937.409289] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ec9faf07-941c-48e2-b95c-86af92cb4576 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.416283] env[63531]: DEBUG oslo_vmware.api [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Waiting for the task: (returnval){ [ 937.416283] env[63531]: value = "task-1118227" [ 937.416283] env[63531]: _type = "Task" [ 937.416283] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.424307] env[63531]: DEBUG oslo_vmware.api [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Task: {'id': task-1118227, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.577957] env[63531]: DEBUG oslo_vmware.api [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118223, 'name': ReconfigVM_Task, 'duration_secs': 0.844876} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.578286] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3fb0499d-8074-4e63-be9f-380730416cc4] Reconfigured VM instance instance-00000052 to attach disk [datastore1] 3fb0499d-8074-4e63-be9f-380730416cc4/3fb0499d-8074-4e63-be9f-380730416cc4.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 937.578950] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f5498b4c-bae7-4f27-850f-91d5b15ac35f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.587182] env[63531]: DEBUG oslo_vmware.api [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 937.587182] env[63531]: value = "task-1118228" [ 937.587182] env[63531]: _type = "Task" [ 937.587182] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.600027] env[63531]: DEBUG oslo_vmware.api [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118228, 'name': Rename_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.648526] env[63531]: DEBUG oslo_vmware.api [None req-ab94488a-b9cf-440f-93f3-139b42979e7b tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118224, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.760649] env[63531]: DEBUG nova.compute.manager [req-b6e71f13-eeb5-4560-b51d-93d2589b3c39 req-fc8ee8f0-d214-4bd8-95bf-d14a72c7082c service nova] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Received event network-vif-plugged-a755a88f-c86c-4183-b9fa-a6133060ea5e {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 937.760870] env[63531]: DEBUG oslo_concurrency.lockutils [req-b6e71f13-eeb5-4560-b51d-93d2589b3c39 req-fc8ee8f0-d214-4bd8-95bf-d14a72c7082c service nova] Acquiring lock "27c23b7c-a68f-43c9-a125-bd657feb3c5b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.763061] env[63531]: DEBUG oslo_concurrency.lockutils [req-b6e71f13-eeb5-4560-b51d-93d2589b3c39 req-fc8ee8f0-d214-4bd8-95bf-d14a72c7082c service nova] Lock "27c23b7c-a68f-43c9-a125-bd657feb3c5b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.763061] env[63531]: DEBUG oslo_concurrency.lockutils [req-b6e71f13-eeb5-4560-b51d-93d2589b3c39 req-fc8ee8f0-d214-4bd8-95bf-d14a72c7082c service nova] Lock "27c23b7c-a68f-43c9-a125-bd657feb3c5b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.763061] env[63531]: DEBUG nova.compute.manager [req-b6e71f13-eeb5-4560-b51d-93d2589b3c39 req-fc8ee8f0-d214-4bd8-95bf-d14a72c7082c service nova] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] No waiting events found dispatching network-vif-plugged-a755a88f-c86c-4183-b9fa-a6133060ea5e {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 937.763061] env[63531]: WARNING nova.compute.manager [req-b6e71f13-eeb5-4560-b51d-93d2589b3c39 req-fc8ee8f0-d214-4bd8-95bf-d14a72c7082c service nova] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Received unexpected event network-vif-plugged-a755a88f-c86c-4183-b9fa-a6133060ea5e for instance with vm_state building and task_state spawning. [ 937.763061] env[63531]: DEBUG nova.compute.manager [req-b6e71f13-eeb5-4560-b51d-93d2589b3c39 req-fc8ee8f0-d214-4bd8-95bf-d14a72c7082c service nova] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Received event network-changed-a755a88f-c86c-4183-b9fa-a6133060ea5e {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 937.763061] env[63531]: DEBUG nova.compute.manager [req-b6e71f13-eeb5-4560-b51d-93d2589b3c39 req-fc8ee8f0-d214-4bd8-95bf-d14a72c7082c service nova] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Refreshing instance network info cache due to event network-changed-a755a88f-c86c-4183-b9fa-a6133060ea5e. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 937.763061] env[63531]: DEBUG oslo_concurrency.lockutils [req-b6e71f13-eeb5-4560-b51d-93d2589b3c39 req-fc8ee8f0-d214-4bd8-95bf-d14a72c7082c service nova] Acquiring lock "refresh_cache-27c23b7c-a68f-43c9-a125-bd657feb3c5b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.763061] env[63531]: DEBUG oslo_concurrency.lockutils [req-b6e71f13-eeb5-4560-b51d-93d2589b3c39 req-fc8ee8f0-d214-4bd8-95bf-d14a72c7082c service nova] Acquired lock "refresh_cache-27c23b7c-a68f-43c9-a125-bd657feb3c5b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.763061] env[63531]: DEBUG nova.network.neutron [req-b6e71f13-eeb5-4560-b51d-93d2589b3c39 req-fc8ee8f0-d214-4bd8-95bf-d14a72c7082c service nova] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Refreshing network info cache for port a755a88f-c86c-4183-b9fa-a6133060ea5e {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 937.771376] env[63531]: DEBUG nova.compute.utils [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 937.780815] env[63531]: DEBUG nova.compute.manager [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] [instance: 55111b87-1e9a-4877-8689-987faa72a54e] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 937.785804] env[63531]: DEBUG nova.network.neutron [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] [instance: 55111b87-1e9a-4877-8689-987faa72a54e] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 937.808488] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118226, 'name': CreateVM_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.847403] env[63531]: DEBUG nova.policy [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ae5c42ec5a444453ae85741e4b56e58d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a06b0b760cfb431d89f488c605edf5db', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 937.934808] env[63531]: DEBUG oslo_vmware.api [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Task: {'id': task-1118227, 'name': PowerOnVM_Task} progress is 94%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.076346] env[63531]: DEBUG nova.network.neutron [req-e5b432cd-0fea-4d4e-91e1-e023fcbca757 req-8249d2de-c9c0-42ce-b0d7-0f6bf498d028 service nova] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Updated VIF entry in instance network info cache for port 1d5fea6d-10c4-4740-81a1-8e1ac927dad5. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 938.076837] env[63531]: DEBUG nova.network.neutron [req-e5b432cd-0fea-4d4e-91e1-e023fcbca757 req-8249d2de-c9c0-42ce-b0d7-0f6bf498d028 service nova] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Updating instance_info_cache with network_info: [{"id": "17da826c-920c-4044-93cb-be8c8cf01c13", "address": "fa:16:3e:9a:ac:ad", "network": {"id": "caf86428-59a5-4e50-9135-2ab8473af64d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1673994696-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.214", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ba8b284854242c392aec5326e996239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap17da826c-92", "ovs_interfaceid": "17da826c-920c-4044-93cb-be8c8cf01c13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "1d5fea6d-10c4-4740-81a1-8e1ac927dad5", "address": "fa:16:3e:06:1e:fa", "network": {"id": "caf86428-59a5-4e50-9135-2ab8473af64d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1673994696-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ba8b284854242c392aec5326e996239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d5fea6d-10", "ovs_interfaceid": "1d5fea6d-10c4-4740-81a1-8e1ac927dad5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 938.098263] env[63531]: DEBUG oslo_vmware.api [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118228, 'name': Rename_Task, 'duration_secs': 0.246627} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.101182] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3fb0499d-8074-4e63-be9f-380730416cc4] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 938.102046] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2243f80c-0c11-46c7-bdd0-8e29ca146106 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.109286] env[63531]: DEBUG oslo_vmware.api [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 938.109286] env[63531]: value = "task-1118229" [ 938.109286] env[63531]: _type = "Task" [ 938.109286] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.117549] env[63531]: DEBUG oslo_vmware.api [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118229, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.133804] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-362619cc-1966-4d27-b307-a70f9831c818 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.145323] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75cefa02-480d-424a-aa01-7bdf5c8ef032 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.154235] env[63531]: DEBUG oslo_vmware.api [None req-ab94488a-b9cf-440f-93f3-139b42979e7b tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118224, 'name': ReconfigVM_Task, 'duration_secs': 1.090344} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.182279] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ab94488a-b9cf-440f-93f3-139b42979e7b tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Releasing lock "a090135f-489d-457d-be36-ba6f61b71ab8" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.182532] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-ab94488a-b9cf-440f-93f3-139b42979e7b tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Reconfigured VM to attach interface {{(pid=63531) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 938.186388] env[63531]: DEBUG nova.network.neutron [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] [instance: 55111b87-1e9a-4877-8689-987faa72a54e] Successfully created port: c2e8b778-965b-40b8-a5c7-8536e055435e {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 938.190303] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fb38f59-35a8-4c7a-b583-984272460d64 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.197679] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2af7fac-3d3a-4bfe-9e0f-80ebfac4448d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.217131] env[63531]: DEBUG nova.compute.provider_tree [None req-4fccfddd-dc56-46bb-aba7-1e6242d29900 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 938.277943] env[63531]: DEBUG nova.compute.manager [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] [instance: 55111b87-1e9a-4877-8689-987faa72a54e] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 938.301425] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118226, 'name': CreateVM_Task, 'duration_secs': 0.581773} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.301611] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 938.302358] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.302523] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.302857] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 938.303130] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eca19597-2c70-458a-85fc-9ee1a6f028a4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.308324] env[63531]: DEBUG oslo_vmware.api [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 938.308324] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52264cc5-4498-eef5-3632-86b7aec65cb0" [ 938.308324] env[63531]: _type = "Task" [ 938.308324] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.318932] env[63531]: DEBUG oslo_vmware.api [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52264cc5-4498-eef5-3632-86b7aec65cb0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.429898] env[63531]: DEBUG oslo_vmware.api [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Task: {'id': task-1118227, 'name': PowerOnVM_Task, 'duration_secs': 0.837558} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.430220] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 938.430381] env[63531]: DEBUG nova.compute.manager [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 938.431233] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41e437e2-87b1-4cb0-bfab-491e60dda45e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.579523] env[63531]: DEBUG oslo_concurrency.lockutils [req-e5b432cd-0fea-4d4e-91e1-e023fcbca757 req-8249d2de-c9c0-42ce-b0d7-0f6bf498d028 service nova] Releasing lock "refresh_cache-a090135f-489d-457d-be36-ba6f61b71ab8" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.583724] env[63531]: DEBUG nova.network.neutron [req-b6e71f13-eeb5-4560-b51d-93d2589b3c39 req-fc8ee8f0-d214-4bd8-95bf-d14a72c7082c service nova] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Updated VIF entry in instance network info cache for port a755a88f-c86c-4183-b9fa-a6133060ea5e. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 938.583828] env[63531]: DEBUG nova.network.neutron [req-b6e71f13-eeb5-4560-b51d-93d2589b3c39 req-fc8ee8f0-d214-4bd8-95bf-d14a72c7082c service nova] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Updating instance_info_cache with network_info: [{"id": "a755a88f-c86c-4183-b9fa-a6133060ea5e", "address": "fa:16:3e:26:e6:a5", "network": {"id": "bf99b8cb-9efb-4a74-9db1-0f5f0875f26a", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-998205595-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2e99349303e413b9f7896d449cb7dff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc6e6fe1-c4f5-4389-a49f-0978060eebb4", "external-id": "nsx-vlan-transportzone-829", "segmentation_id": 829, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa755a88f-c8", "ovs_interfaceid": "a755a88f-c86c-4183-b9fa-a6133060ea5e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 938.619723] env[63531]: DEBUG oslo_vmware.api [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118229, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.687443] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ab94488a-b9cf-440f-93f3-139b42979e7b tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "interface-a090135f-489d-457d-be36-ba6f61b71ab8-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.649s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.721673] env[63531]: DEBUG nova.scheduler.client.report [None req-4fccfddd-dc56-46bb-aba7-1e6242d29900 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 938.819635] env[63531]: DEBUG oslo_vmware.api [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52264cc5-4498-eef5-3632-86b7aec65cb0, 'name': SearchDatastore_Task, 'duration_secs': 0.017915} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.820228] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.820361] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 938.820653] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.820814] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.821012] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 938.821328] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8359ca51-baef-4243-89e0-bb6e7dc55db7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.829781] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 938.829967] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 938.830713] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8ff8a33c-7bdc-439c-8615-35e05a54da09 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.836270] env[63531]: DEBUG oslo_vmware.api [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 938.836270] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52df7b10-e1ef-87c6-fc1f-972ec1091a83" [ 938.836270] env[63531]: _type = "Task" [ 938.836270] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.844413] env[63531]: DEBUG oslo_vmware.api [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52df7b10-e1ef-87c6-fc1f-972ec1091a83, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.948370] env[63531]: DEBUG oslo_concurrency.lockutils [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.088696] env[63531]: DEBUG oslo_concurrency.lockutils [req-b6e71f13-eeb5-4560-b51d-93d2589b3c39 req-fc8ee8f0-d214-4bd8-95bf-d14a72c7082c service nova] Releasing lock "refresh_cache-27c23b7c-a68f-43c9-a125-bd657feb3c5b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.123342] env[63531]: DEBUG oslo_vmware.api [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118229, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.227638] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4fccfddd-dc56-46bb-aba7-1e6242d29900 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.960s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.231040] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4b9a0fd3-15d1-440d-90b4-9ce5235e55e0 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.771s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.231040] env[63531]: DEBUG nova.objects.instance [None req-4b9a0fd3-15d1-440d-90b4-9ce5235e55e0 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lazy-loading 'resources' on Instance uuid 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 939.232479] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4c36b677-1dba-4034-8c3f-d7d2eda1da73 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Acquiring lock "e065f052-ba3e-4783-8953-5dc200d1f3e9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.232824] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4c36b677-1dba-4034-8c3f-d7d2eda1da73 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Lock "e065f052-ba3e-4783-8953-5dc200d1f3e9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.233152] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4c36b677-1dba-4034-8c3f-d7d2eda1da73 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Acquiring lock "e065f052-ba3e-4783-8953-5dc200d1f3e9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.233386] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4c36b677-1dba-4034-8c3f-d7d2eda1da73 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Lock "e065f052-ba3e-4783-8953-5dc200d1f3e9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.233665] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4c36b677-1dba-4034-8c3f-d7d2eda1da73 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Lock "e065f052-ba3e-4783-8953-5dc200d1f3e9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.235928] env[63531]: INFO nova.compute.manager [None req-4c36b677-1dba-4034-8c3f-d7d2eda1da73 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Terminating instance [ 939.237920] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4c36b677-1dba-4034-8c3f-d7d2eda1da73 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Acquiring lock "refresh_cache-e065f052-ba3e-4783-8953-5dc200d1f3e9" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.238221] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4c36b677-1dba-4034-8c3f-d7d2eda1da73 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Acquired lock "refresh_cache-e065f052-ba3e-4783-8953-5dc200d1f3e9" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.238448] env[63531]: DEBUG nova.network.neutron [None req-4c36b677-1dba-4034-8c3f-d7d2eda1da73 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 939.255187] env[63531]: INFO nova.scheduler.client.report [None req-4fccfddd-dc56-46bb-aba7-1e6242d29900 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Deleted allocations for instance 06ec91b2-14cb-4db8-8fab-e6c4183b1116 [ 939.289093] env[63531]: DEBUG nova.compute.manager [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] [instance: 55111b87-1e9a-4877-8689-987faa72a54e] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 939.314988] env[63531]: DEBUG nova.virt.hardware [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 939.315276] env[63531]: DEBUG nova.virt.hardware [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 939.315438] env[63531]: DEBUG nova.virt.hardware [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 939.315627] env[63531]: DEBUG nova.virt.hardware [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 939.315781] env[63531]: DEBUG nova.virt.hardware [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 939.315934] env[63531]: DEBUG nova.virt.hardware [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 939.316161] env[63531]: DEBUG nova.virt.hardware [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 939.316326] env[63531]: DEBUG nova.virt.hardware [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 939.316498] env[63531]: DEBUG nova.virt.hardware [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 939.316666] env[63531]: DEBUG nova.virt.hardware [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 939.316859] env[63531]: DEBUG nova.virt.hardware [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 939.318013] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9a89f19-22fa-49e9-a25f-f7104973afe9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.325903] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa5d69c1-857d-41d0-b49f-c46ef42643a9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.348527] env[63531]: DEBUG oslo_vmware.api [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52df7b10-e1ef-87c6-fc1f-972ec1091a83, 'name': SearchDatastore_Task, 'duration_secs': 0.011177} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.350136] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ac23964f-eb56-43ad-93ba-59618b8d1ecf {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.353419] env[63531]: DEBUG nova.compute.manager [req-9aa4eecb-2c2e-485b-b887-f5aba0d969d1 req-e0da7496-b2f1-4ce3-bbcc-3f8b666dddbd service nova] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Received event network-changed-7433ce46-ff5a-4e78-bcc9-9dcf2264a01b {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 939.353607] env[63531]: DEBUG nova.compute.manager [req-9aa4eecb-2c2e-485b-b887-f5aba0d969d1 req-e0da7496-b2f1-4ce3-bbcc-3f8b666dddbd service nova] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Refreshing instance network info cache due to event network-changed-7433ce46-ff5a-4e78-bcc9-9dcf2264a01b. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 939.353826] env[63531]: DEBUG oslo_concurrency.lockutils [req-9aa4eecb-2c2e-485b-b887-f5aba0d969d1 req-e0da7496-b2f1-4ce3-bbcc-3f8b666dddbd service nova] Acquiring lock "refresh_cache-ccf00b09-29fe-4a76-a8af-97774acd77de" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.353963] env[63531]: DEBUG oslo_concurrency.lockutils [req-9aa4eecb-2c2e-485b-b887-f5aba0d969d1 req-e0da7496-b2f1-4ce3-bbcc-3f8b666dddbd service nova] Acquired lock "refresh_cache-ccf00b09-29fe-4a76-a8af-97774acd77de" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.354143] env[63531]: DEBUG nova.network.neutron [req-9aa4eecb-2c2e-485b-b887-f5aba0d969d1 req-e0da7496-b2f1-4ce3-bbcc-3f8b666dddbd service nova] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Refreshing network info cache for port 7433ce46-ff5a-4e78-bcc9-9dcf2264a01b {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 939.359163] env[63531]: DEBUG oslo_vmware.api [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 939.359163] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52352e9c-fab5-64a6-400f-5aa46d21562d" [ 939.359163] env[63531]: _type = "Task" [ 939.359163] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.367623] env[63531]: DEBUG oslo_vmware.api [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52352e9c-fab5-64a6-400f-5aa46d21562d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.620207] env[63531]: DEBUG oslo_vmware.api [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118229, 'name': PowerOnVM_Task, 'duration_secs': 1.05124} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.620493] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3fb0499d-8074-4e63-be9f-380730416cc4] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 939.620711] env[63531]: INFO nova.compute.manager [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3fb0499d-8074-4e63-be9f-380730416cc4] Took 7.70 seconds to spawn the instance on the hypervisor. [ 939.620895] env[63531]: DEBUG nova.compute.manager [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3fb0499d-8074-4e63-be9f-380730416cc4] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 939.621681] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8a2cbb3-d1d4-418c-a262-e856862498fc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.761132] env[63531]: DEBUG nova.network.neutron [None req-4c36b677-1dba-4034-8c3f-d7d2eda1da73 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 939.763243] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4fccfddd-dc56-46bb-aba7-1e6242d29900 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "06ec91b2-14cb-4db8-8fab-e6c4183b1116" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.106s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.804193] env[63531]: DEBUG nova.compute.manager [req-e3992a2f-acc1-4cc4-a1d4-61dc266a716b req-0d06ca07-125e-4a17-bc84-2520343d07da service nova] [instance: 55111b87-1e9a-4877-8689-987faa72a54e] Received event network-vif-plugged-c2e8b778-965b-40b8-a5c7-8536e055435e {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 939.804556] env[63531]: DEBUG oslo_concurrency.lockutils [req-e3992a2f-acc1-4cc4-a1d4-61dc266a716b req-0d06ca07-125e-4a17-bc84-2520343d07da service nova] Acquiring lock "55111b87-1e9a-4877-8689-987faa72a54e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.804851] env[63531]: DEBUG oslo_concurrency.lockutils [req-e3992a2f-acc1-4cc4-a1d4-61dc266a716b req-0d06ca07-125e-4a17-bc84-2520343d07da service nova] Lock "55111b87-1e9a-4877-8689-987faa72a54e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.805137] env[63531]: DEBUG oslo_concurrency.lockutils [req-e3992a2f-acc1-4cc4-a1d4-61dc266a716b req-0d06ca07-125e-4a17-bc84-2520343d07da service nova] Lock "55111b87-1e9a-4877-8689-987faa72a54e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.805384] env[63531]: DEBUG nova.compute.manager [req-e3992a2f-acc1-4cc4-a1d4-61dc266a716b req-0d06ca07-125e-4a17-bc84-2520343d07da service nova] [instance: 55111b87-1e9a-4877-8689-987faa72a54e] No waiting events found dispatching network-vif-plugged-c2e8b778-965b-40b8-a5c7-8536e055435e {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 939.805626] env[63531]: WARNING nova.compute.manager [req-e3992a2f-acc1-4cc4-a1d4-61dc266a716b req-0d06ca07-125e-4a17-bc84-2520343d07da service nova] [instance: 55111b87-1e9a-4877-8689-987faa72a54e] Received unexpected event network-vif-plugged-c2e8b778-965b-40b8-a5c7-8536e055435e for instance with vm_state building and task_state spawning. [ 939.836879] env[63531]: DEBUG nova.network.neutron [None req-4c36b677-1dba-4034-8c3f-d7d2eda1da73 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 939.860376] env[63531]: DEBUG nova.network.neutron [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] [instance: 55111b87-1e9a-4877-8689-987faa72a54e] Successfully updated port: c2e8b778-965b-40b8-a5c7-8536e055435e {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 939.874051] env[63531]: DEBUG oslo_vmware.api [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52352e9c-fab5-64a6-400f-5aa46d21562d, 'name': SearchDatastore_Task, 'duration_secs': 0.011323} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.877964] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.877964] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 27c23b7c-a68f-43c9-a125-bd657feb3c5b/27c23b7c-a68f-43c9-a125-bd657feb3c5b.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 939.880191] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c676c196-1d04-48b8-8983-a6db1f70b43c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.889932] env[63531]: DEBUG oslo_vmware.api [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 939.889932] env[63531]: value = "task-1118230" [ 939.889932] env[63531]: _type = "Task" [ 939.889932] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.904164] env[63531]: DEBUG oslo_vmware.api [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118230, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.054349] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd735bf4-2de1-4e60-b4c2-0d44c8b836d6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.061915] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11cfb324-fbb8-4746-b75a-0533a6996c5c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.096180] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b6a8569-53af-4b8d-93e9-3bce9a7ed1ed {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.104543] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9357a51d-3ee3-43d8-af10-f296d1e3486c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.118837] env[63531]: DEBUG nova.compute.provider_tree [None req-4b9a0fd3-15d1-440d-90b4-9ce5235e55e0 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 940.140964] env[63531]: INFO nova.compute.manager [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3fb0499d-8074-4e63-be9f-380730416cc4] Took 34.01 seconds to build instance. [ 940.340986] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4c36b677-1dba-4034-8c3f-d7d2eda1da73 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Releasing lock "refresh_cache-e065f052-ba3e-4783-8953-5dc200d1f3e9" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 940.340986] env[63531]: DEBUG nova.compute.manager [None req-4c36b677-1dba-4034-8c3f-d7d2eda1da73 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 940.341400] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-4c36b677-1dba-4034-8c3f-d7d2eda1da73 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 940.342540] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4eeb2851-bfe5-41c1-9789-5dc0636db8f9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.353367] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c36b677-1dba-4034-8c3f-d7d2eda1da73 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 940.353699] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6a562b30-c713-4fe1-86c7-4bfe6bf24a3e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.362136] env[63531]: DEBUG oslo_vmware.api [None req-4c36b677-1dba-4034-8c3f-d7d2eda1da73 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Waiting for the task: (returnval){ [ 940.362136] env[63531]: value = "task-1118231" [ 940.362136] env[63531]: _type = "Task" [ 940.362136] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.369412] env[63531]: DEBUG oslo_concurrency.lockutils [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Acquiring lock "refresh_cache-55111b87-1e9a-4877-8689-987faa72a54e" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 940.369871] env[63531]: DEBUG oslo_concurrency.lockutils [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Acquired lock "refresh_cache-55111b87-1e9a-4877-8689-987faa72a54e" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.369871] env[63531]: DEBUG nova.network.neutron [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] [instance: 55111b87-1e9a-4877-8689-987faa72a54e] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 940.382672] env[63531]: DEBUG oslo_vmware.api [None req-4c36b677-1dba-4034-8c3f-d7d2eda1da73 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Task: {'id': task-1118231, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.385973] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3ac1abf3-9cc3-459d-89d6-e0be3bbf01b0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "interface-a090135f-489d-457d-be36-ba6f61b71ab8-1d5fea6d-10c4-4740-81a1-8e1ac927dad5" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.386327] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3ac1abf3-9cc3-459d-89d6-e0be3bbf01b0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "interface-a090135f-489d-457d-be36-ba6f61b71ab8-1d5fea6d-10c4-4740-81a1-8e1ac927dad5" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.408425] env[63531]: DEBUG oslo_vmware.api [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118230, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.431083] env[63531]: DEBUG nova.network.neutron [req-9aa4eecb-2c2e-485b-b887-f5aba0d969d1 req-e0da7496-b2f1-4ce3-bbcc-3f8b666dddbd service nova] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Updated VIF entry in instance network info cache for port 7433ce46-ff5a-4e78-bcc9-9dcf2264a01b. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 940.431713] env[63531]: DEBUG nova.network.neutron [req-9aa4eecb-2c2e-485b-b887-f5aba0d969d1 req-e0da7496-b2f1-4ce3-bbcc-3f8b666dddbd service nova] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Updating instance_info_cache with network_info: [{"id": "7433ce46-ff5a-4e78-bcc9-9dcf2264a01b", "address": "fa:16:3e:ad:77:b5", "network": {"id": "913045d9-e90b-4366-987c-7706b73a6147", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1696510389-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f71c7db0a3614995becc25d9a8dce787", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c297fe21-cd0b-4226-813b-a65d2358d034", "external-id": "nsx-vlan-transportzone-98", "segmentation_id": 98, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7433ce46-ff", "ovs_interfaceid": "7433ce46-ff5a-4e78-bcc9-9dcf2264a01b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.623275] env[63531]: DEBUG nova.scheduler.client.report [None req-4b9a0fd3-15d1-440d-90b4-9ce5235e55e0 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 940.643740] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7c19c69b-dba8-416e-867d-535ce0437b11 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "3fb0499d-8074-4e63-be9f-380730416cc4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.529s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.874542] env[63531]: DEBUG oslo_vmware.api [None req-4c36b677-1dba-4034-8c3f-d7d2eda1da73 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Task: {'id': task-1118231, 'name': PowerOffVM_Task, 'duration_secs': 0.153333} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.874542] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c36b677-1dba-4034-8c3f-d7d2eda1da73 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 940.874542] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-4c36b677-1dba-4034-8c3f-d7d2eda1da73 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 940.875226] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0f7aa9bd-8786-48fe-a245-72cd18ac573e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.889142] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3ac1abf3-9cc3-459d-89d6-e0be3bbf01b0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "a090135f-489d-457d-be36-ba6f61b71ab8" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 940.889788] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3ac1abf3-9cc3-459d-89d6-e0be3bbf01b0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquired lock "a090135f-489d-457d-be36-ba6f61b71ab8" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.890179] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0415e88-2c2e-40e1-bab4-3eac8033653b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.916107] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e46902bd-4224-4bbd-a370-98008e322ce2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.918842] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-4c36b677-1dba-4034-8c3f-d7d2eda1da73 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 940.919091] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-4c36b677-1dba-4034-8c3f-d7d2eda1da73 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 940.919287] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-4c36b677-1dba-4034-8c3f-d7d2eda1da73 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Deleting the datastore file [datastore1] e065f052-ba3e-4783-8953-5dc200d1f3e9 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 940.919565] env[63531]: DEBUG oslo_vmware.api [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118230, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.591526} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.919786] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-669b16ab-6064-406a-9e03-f4618191f7c1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.922098] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 27c23b7c-a68f-43c9-a125-bd657feb3c5b/27c23b7c-a68f-43c9-a125-bd657feb3c5b.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 940.922814] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 940.923605] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-283928af-453e-41de-bf5b-efbcf3c87e5c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.945153] env[63531]: DEBUG oslo_concurrency.lockutils [req-9aa4eecb-2c2e-485b-b887-f5aba0d969d1 req-e0da7496-b2f1-4ce3-bbcc-3f8b666dddbd service nova] Releasing lock "refresh_cache-ccf00b09-29fe-4a76-a8af-97774acd77de" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 940.951193] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-3ac1abf3-9cc3-459d-89d6-e0be3bbf01b0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Reconfiguring VM to detach interface {{(pid=63531) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 940.953316] env[63531]: DEBUG nova.network.neutron [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] [instance: 55111b87-1e9a-4877-8689-987faa72a54e] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 940.955519] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-158ade2a-bdb7-4061-aaf8-3b6bb908f900 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.968787] env[63531]: DEBUG oslo_vmware.api [None req-4c36b677-1dba-4034-8c3f-d7d2eda1da73 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Waiting for the task: (returnval){ [ 940.968787] env[63531]: value = "task-1118233" [ 940.968787] env[63531]: _type = "Task" [ 940.968787] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.971891] env[63531]: DEBUG oslo_vmware.api [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 940.971891] env[63531]: value = "task-1118234" [ 940.971891] env[63531]: _type = "Task" [ 940.971891] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.983215] env[63531]: DEBUG oslo_vmware.api [None req-3ac1abf3-9cc3-459d-89d6-e0be3bbf01b0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for the task: (returnval){ [ 940.983215] env[63531]: value = "task-1118235" [ 940.983215] env[63531]: _type = "Task" [ 940.983215] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.992520] env[63531]: DEBUG oslo_vmware.api [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118234, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.992781] env[63531]: DEBUG oslo_vmware.api [None req-4c36b677-1dba-4034-8c3f-d7d2eda1da73 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Task: {'id': task-1118233, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.999471] env[63531]: DEBUG oslo_vmware.api [None req-3ac1abf3-9cc3-459d-89d6-e0be3bbf01b0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118235, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.130388] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4b9a0fd3-15d1-440d-90b4-9ce5235e55e0 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.900s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.133420] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.098s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.134526] env[63531]: INFO nova.compute.claims [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 941.159841] env[63531]: INFO nova.scheduler.client.report [None req-4b9a0fd3-15d1-440d-90b4-9ce5235e55e0 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Deleted allocations for instance 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9 [ 941.230875] env[63531]: DEBUG nova.network.neutron [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] [instance: 55111b87-1e9a-4877-8689-987faa72a54e] Updating instance_info_cache with network_info: [{"id": "c2e8b778-965b-40b8-a5c7-8536e055435e", "address": "fa:16:3e:b3:64:d0", "network": {"id": "0ab1224b-f3f0-405a-9eae-bad15386c616", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-1258356972-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a06b0b760cfb431d89f488c605edf5db", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bec1528b-3e87-477b-8ab2-02696ad47e66", "external-id": "nsx-vlan-transportzone-180", "segmentation_id": 180, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2e8b778-96", "ovs_interfaceid": "c2e8b778-965b-40b8-a5c7-8536e055435e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 941.276265] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.276524] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.481771] env[63531]: DEBUG oslo_vmware.api [None req-4c36b677-1dba-4034-8c3f-d7d2eda1da73 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Task: {'id': task-1118233, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.42585} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.482576] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-4c36b677-1dba-4034-8c3f-d7d2eda1da73 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 941.482798] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-4c36b677-1dba-4034-8c3f-d7d2eda1da73 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 941.483017] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-4c36b677-1dba-4034-8c3f-d7d2eda1da73 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 941.483214] env[63531]: INFO nova.compute.manager [None req-4c36b677-1dba-4034-8c3f-d7d2eda1da73 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Took 1.14 seconds to destroy the instance on the hypervisor. [ 941.483451] env[63531]: DEBUG oslo.service.loopingcall [None req-4c36b677-1dba-4034-8c3f-d7d2eda1da73 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 941.483667] env[63531]: DEBUG nova.compute.manager [-] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 941.483786] env[63531]: DEBUG nova.network.neutron [-] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 941.492021] env[63531]: DEBUG oslo_vmware.api [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118234, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.11353} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.492753] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 941.493669] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9440b55c-5fcb-48cf-a7f5-26554f36fa2d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.501594] env[63531]: DEBUG oslo_vmware.api [None req-3ac1abf3-9cc3-459d-89d6-e0be3bbf01b0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118235, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.521119] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Reconfiguring VM instance instance-00000053 to attach disk [datastore1] 27c23b7c-a68f-43c9-a125-bd657feb3c5b/27c23b7c-a68f-43c9-a125-bd657feb3c5b.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 941.521930] env[63531]: DEBUG nova.network.neutron [-] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 941.523131] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d17e8752-bf4d-4e11-b785-78a0a2add768 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.543923] env[63531]: DEBUG oslo_vmware.api [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 941.543923] env[63531]: value = "task-1118236" [ 941.543923] env[63531]: _type = "Task" [ 941.543923] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.551969] env[63531]: DEBUG oslo_vmware.api [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118236, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.667695] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4b9a0fd3-15d1-440d-90b4-9ce5235e55e0 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.415s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.737417] env[63531]: DEBUG oslo_concurrency.lockutils [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Releasing lock "refresh_cache-55111b87-1e9a-4877-8689-987faa72a54e" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.737905] env[63531]: DEBUG nova.compute.manager [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] [instance: 55111b87-1e9a-4877-8689-987faa72a54e] Instance network_info: |[{"id": "c2e8b778-965b-40b8-a5c7-8536e055435e", "address": "fa:16:3e:b3:64:d0", "network": {"id": "0ab1224b-f3f0-405a-9eae-bad15386c616", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-1258356972-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a06b0b760cfb431d89f488c605edf5db", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bec1528b-3e87-477b-8ab2-02696ad47e66", "external-id": "nsx-vlan-transportzone-180", "segmentation_id": 180, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2e8b778-96", "ovs_interfaceid": "c2e8b778-965b-40b8-a5c7-8536e055435e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 941.738441] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] [instance: 55111b87-1e9a-4877-8689-987faa72a54e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b3:64:d0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bec1528b-3e87-477b-8ab2-02696ad47e66', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c2e8b778-965b-40b8-a5c7-8536e055435e', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 941.747278] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Creating folder: Project (a06b0b760cfb431d89f488c605edf5db). Parent ref: group-v244585. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 941.747637] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d455d4fb-ed32-4e53-b94b-d1e1ca4fbd25 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.759485] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Created folder: Project (a06b0b760cfb431d89f488c605edf5db) in parent group-v244585. [ 941.759679] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Creating folder: Instances. Parent ref: group-v244757. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 941.759922] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-38d864d8-4cc7-415b-9fd6-0e79e03ff0f1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.768719] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Created folder: Instances in parent group-v244757. [ 941.768947] env[63531]: DEBUG oslo.service.loopingcall [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 941.769151] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 55111b87-1e9a-4877-8689-987faa72a54e] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 941.769356] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d80fd5f7-f737-44b5-92c7-a20f6482e9db {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.785912] env[63531]: DEBUG nova.compute.manager [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 941.793181] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 941.793181] env[63531]: value = "task-1118239" [ 941.793181] env[63531]: _type = "Task" [ 941.793181] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.802967] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118239, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.813738] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ba2842a1-0c0c-4cb4-b87b-242607074b48 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "3fb0499d-8074-4e63-be9f-380730416cc4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.814102] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ba2842a1-0c0c-4cb4-b87b-242607074b48 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "3fb0499d-8074-4e63-be9f-380730416cc4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.814571] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ba2842a1-0c0c-4cb4-b87b-242607074b48 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "3fb0499d-8074-4e63-be9f-380730416cc4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.814656] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ba2842a1-0c0c-4cb4-b87b-242607074b48 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "3fb0499d-8074-4e63-be9f-380730416cc4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.814938] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ba2842a1-0c0c-4cb4-b87b-242607074b48 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "3fb0499d-8074-4e63-be9f-380730416cc4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.817720] env[63531]: INFO nova.compute.manager [None req-ba2842a1-0c0c-4cb4-b87b-242607074b48 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3fb0499d-8074-4e63-be9f-380730416cc4] Terminating instance [ 941.822954] env[63531]: DEBUG nova.compute.manager [None req-ba2842a1-0c0c-4cb4-b87b-242607074b48 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3fb0499d-8074-4e63-be9f-380730416cc4] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 941.823176] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-ba2842a1-0c0c-4cb4-b87b-242607074b48 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3fb0499d-8074-4e63-be9f-380730416cc4] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 941.824037] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2cb94a0-706a-4c08-94e5-98f8b3ef4c8b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.831707] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba2842a1-0c0c-4cb4-b87b-242607074b48 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3fb0499d-8074-4e63-be9f-380730416cc4] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 941.833030] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-41f88305-dbe4-430a-9618-e59456fad2e4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.835586] env[63531]: DEBUG nova.compute.manager [req-a4622ad1-054e-4cc6-8594-23f46fec8a15 req-9e651580-a717-41dd-9b38-e94368c5e8b6 service nova] [instance: 55111b87-1e9a-4877-8689-987faa72a54e] Received event network-changed-c2e8b778-965b-40b8-a5c7-8536e055435e {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 941.835783] env[63531]: DEBUG nova.compute.manager [req-a4622ad1-054e-4cc6-8594-23f46fec8a15 req-9e651580-a717-41dd-9b38-e94368c5e8b6 service nova] [instance: 55111b87-1e9a-4877-8689-987faa72a54e] Refreshing instance network info cache due to event network-changed-c2e8b778-965b-40b8-a5c7-8536e055435e. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 941.835997] env[63531]: DEBUG oslo_concurrency.lockutils [req-a4622ad1-054e-4cc6-8594-23f46fec8a15 req-9e651580-a717-41dd-9b38-e94368c5e8b6 service nova] Acquiring lock "refresh_cache-55111b87-1e9a-4877-8689-987faa72a54e" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.836165] env[63531]: DEBUG oslo_concurrency.lockutils [req-a4622ad1-054e-4cc6-8594-23f46fec8a15 req-9e651580-a717-41dd-9b38-e94368c5e8b6 service nova] Acquired lock "refresh_cache-55111b87-1e9a-4877-8689-987faa72a54e" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.836329] env[63531]: DEBUG nova.network.neutron [req-a4622ad1-054e-4cc6-8594-23f46fec8a15 req-9e651580-a717-41dd-9b38-e94368c5e8b6 service nova] [instance: 55111b87-1e9a-4877-8689-987faa72a54e] Refreshing network info cache for port c2e8b778-965b-40b8-a5c7-8536e055435e {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 941.844668] env[63531]: DEBUG oslo_vmware.api [None req-ba2842a1-0c0c-4cb4-b87b-242607074b48 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 941.844668] env[63531]: value = "task-1118240" [ 941.844668] env[63531]: _type = "Task" [ 941.844668] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.853758] env[63531]: DEBUG oslo_vmware.api [None req-ba2842a1-0c0c-4cb4-b87b-242607074b48 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118240, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.999342] env[63531]: DEBUG oslo_vmware.api [None req-3ac1abf3-9cc3-459d-89d6-e0be3bbf01b0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118235, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.039026] env[63531]: DEBUG nova.network.neutron [-] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 942.054326] env[63531]: DEBUG oslo_vmware.api [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118236, 'name': ReconfigVM_Task, 'duration_secs': 0.458388} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.054706] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Reconfigured VM instance instance-00000053 to attach disk [datastore1] 27c23b7c-a68f-43c9-a125-bd657feb3c5b/27c23b7c-a68f-43c9-a125-bd657feb3c5b.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 942.055329] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cecbfa78-7250-452d-bb60-e0d8f5fe63db {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.060904] env[63531]: DEBUG oslo_vmware.api [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 942.060904] env[63531]: value = "task-1118241" [ 942.060904] env[63531]: _type = "Task" [ 942.060904] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.069752] env[63531]: DEBUG oslo_vmware.api [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118241, 'name': Rename_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.306149] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118239, 'name': CreateVM_Task, 'duration_secs': 0.318877} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.307086] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.307306] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 55111b87-1e9a-4877-8689-987faa72a54e] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 942.307997] env[63531]: DEBUG oslo_concurrency.lockutils [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.308985] env[63531]: DEBUG oslo_concurrency.lockutils [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.308985] env[63531]: DEBUG oslo_concurrency.lockutils [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 942.310877] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-463b95fd-8d57-44f5-96e6-7905f9f55b79 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.315543] env[63531]: DEBUG oslo_vmware.api [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Waiting for the task: (returnval){ [ 942.315543] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]524d589e-884c-e768-6663-b2464fcb1cd2" [ 942.315543] env[63531]: _type = "Task" [ 942.315543] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.323457] env[63531]: DEBUG oslo_vmware.api [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]524d589e-884c-e768-6663-b2464fcb1cd2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.356568] env[63531]: DEBUG oslo_vmware.api [None req-ba2842a1-0c0c-4cb4-b87b-242607074b48 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118240, 'name': PowerOffVM_Task, 'duration_secs': 0.223256} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.359044] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba2842a1-0c0c-4cb4-b87b-242607074b48 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3fb0499d-8074-4e63-be9f-380730416cc4] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 942.359250] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-ba2842a1-0c0c-4cb4-b87b-242607074b48 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3fb0499d-8074-4e63-be9f-380730416cc4] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 942.360051] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2e27a680-1775-415e-9847-deb13e95451b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.410793] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ceb2f12-0879-4f06-a878-2a332c085483 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.418137] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d984a0b-db94-44dd-b566-e76bf4da754b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.449337] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cf148fe-ba74-4576-a9b0-14cbba609b59 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.457516] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af3ecbb8-b0e8-4a19-822f-36582842250e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.461342] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-ba2842a1-0c0c-4cb4-b87b-242607074b48 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3fb0499d-8074-4e63-be9f-380730416cc4] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 942.461550] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-ba2842a1-0c0c-4cb4-b87b-242607074b48 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3fb0499d-8074-4e63-be9f-380730416cc4] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 942.461729] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba2842a1-0c0c-4cb4-b87b-242607074b48 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Deleting the datastore file [datastore1] 3fb0499d-8074-4e63-be9f-380730416cc4 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 942.464254] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b88dd227-6228-43f1-b42d-be3e30668ab3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.477500] env[63531]: DEBUG nova.compute.provider_tree [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 942.480463] env[63531]: DEBUG oslo_vmware.api [None req-ba2842a1-0c0c-4cb4-b87b-242607074b48 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 942.480463] env[63531]: value = "task-1118243" [ 942.480463] env[63531]: _type = "Task" [ 942.480463] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.493039] env[63531]: DEBUG oslo_vmware.api [None req-ba2842a1-0c0c-4cb4-b87b-242607074b48 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118243, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.500649] env[63531]: DEBUG oslo_vmware.api [None req-3ac1abf3-9cc3-459d-89d6-e0be3bbf01b0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118235, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.541536] env[63531]: INFO nova.compute.manager [-] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Took 1.06 seconds to deallocate network for instance. [ 942.570809] env[63531]: DEBUG oslo_vmware.api [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118241, 'name': Rename_Task, 'duration_secs': 0.156331} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.571249] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 942.571713] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6ab58fca-70e7-48ca-b786-eadb98e567ae {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.577096] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c8a73bd9-9696-4dd3-baf2-1c46189464fe tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquiring lock "48043b5d-c0d3-4e8a-8d28-5a408d800e11" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.577561] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c8a73bd9-9696-4dd3-baf2-1c46189464fe tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "48043b5d-c0d3-4e8a-8d28-5a408d800e11" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.580702] env[63531]: DEBUG oslo_vmware.api [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 942.580702] env[63531]: value = "task-1118244" [ 942.580702] env[63531]: _type = "Task" [ 942.580702] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.589961] env[63531]: DEBUG oslo_vmware.api [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118244, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.733374] env[63531]: DEBUG nova.network.neutron [req-a4622ad1-054e-4cc6-8594-23f46fec8a15 req-9e651580-a717-41dd-9b38-e94368c5e8b6 service nova] [instance: 55111b87-1e9a-4877-8689-987faa72a54e] Updated VIF entry in instance network info cache for port c2e8b778-965b-40b8-a5c7-8536e055435e. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 942.733740] env[63531]: DEBUG nova.network.neutron [req-a4622ad1-054e-4cc6-8594-23f46fec8a15 req-9e651580-a717-41dd-9b38-e94368c5e8b6 service nova] [instance: 55111b87-1e9a-4877-8689-987faa72a54e] Updating instance_info_cache with network_info: [{"id": "c2e8b778-965b-40b8-a5c7-8536e055435e", "address": "fa:16:3e:b3:64:d0", "network": {"id": "0ab1224b-f3f0-405a-9eae-bad15386c616", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-1258356972-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a06b0b760cfb431d89f488c605edf5db", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bec1528b-3e87-477b-8ab2-02696ad47e66", "external-id": "nsx-vlan-transportzone-180", "segmentation_id": 180, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2e8b778-96", "ovs_interfaceid": "c2e8b778-965b-40b8-a5c7-8536e055435e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 942.826042] env[63531]: DEBUG oslo_vmware.api [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]524d589e-884c-e768-6663-b2464fcb1cd2, 'name': SearchDatastore_Task, 'duration_secs': 0.009737} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.826380] env[63531]: DEBUG oslo_concurrency.lockutils [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 942.826621] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] [instance: 55111b87-1e9a-4877-8689-987faa72a54e] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 942.826869] env[63531]: DEBUG oslo_concurrency.lockutils [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.827046] env[63531]: DEBUG oslo_concurrency.lockutils [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.827251] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 942.827521] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-12bfc83f-d6b5-48fa-99fa-a3eba22f8eba {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.836034] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 942.836238] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 942.836965] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a0e50c4a-d171-4623-86a9-25c8729b5104 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.842080] env[63531]: DEBUG oslo_vmware.api [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Waiting for the task: (returnval){ [ 942.842080] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52a0f35d-1e25-22dd-1271-4c913c072701" [ 942.842080] env[63531]: _type = "Task" [ 942.842080] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.849641] env[63531]: DEBUG oslo_vmware.api [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52a0f35d-1e25-22dd-1271-4c913c072701, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.984690] env[63531]: DEBUG nova.scheduler.client.report [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 943.002847] env[63531]: DEBUG oslo_vmware.api [None req-ba2842a1-0c0c-4cb4-b87b-242607074b48 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118243, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.194369} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.008720] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba2842a1-0c0c-4cb4-b87b-242607074b48 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 943.009054] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-ba2842a1-0c0c-4cb4-b87b-242607074b48 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3fb0499d-8074-4e63-be9f-380730416cc4] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 943.009269] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-ba2842a1-0c0c-4cb4-b87b-242607074b48 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3fb0499d-8074-4e63-be9f-380730416cc4] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 943.009497] env[63531]: INFO nova.compute.manager [None req-ba2842a1-0c0c-4cb4-b87b-242607074b48 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3fb0499d-8074-4e63-be9f-380730416cc4] Took 1.19 seconds to destroy the instance on the hypervisor. [ 943.009807] env[63531]: DEBUG oslo.service.loopingcall [None req-ba2842a1-0c0c-4cb4-b87b-242607074b48 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 943.010482] env[63531]: DEBUG oslo_vmware.api [None req-3ac1abf3-9cc3-459d-89d6-e0be3bbf01b0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118235, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.010782] env[63531]: DEBUG nova.compute.manager [-] [instance: 3fb0499d-8074-4e63-be9f-380730416cc4] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 943.010905] env[63531]: DEBUG nova.network.neutron [-] [instance: 3fb0499d-8074-4e63-be9f-380730416cc4] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 943.048137] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4c36b677-1dba-4034-8c3f-d7d2eda1da73 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.082508] env[63531]: INFO nova.compute.manager [None req-c8a73bd9-9696-4dd3-baf2-1c46189464fe tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Detaching volume 02e72028-7ac8-4eab-88b0-9de83339ce64 [ 943.096174] env[63531]: DEBUG oslo_vmware.api [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118244, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.119743] env[63531]: INFO nova.virt.block_device [None req-c8a73bd9-9696-4dd3-baf2-1c46189464fe tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Attempting to driver detach volume 02e72028-7ac8-4eab-88b0-9de83339ce64 from mountpoint /dev/sdb [ 943.121094] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-c8a73bd9-9696-4dd3-baf2-1c46189464fe tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Volume detach. Driver type: vmdk {{(pid=63531) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 943.121094] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-c8a73bd9-9696-4dd3-baf2-1c46189464fe tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244696', 'volume_id': '02e72028-7ac8-4eab-88b0-9de83339ce64', 'name': 'volume-02e72028-7ac8-4eab-88b0-9de83339ce64', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '48043b5d-c0d3-4e8a-8d28-5a408d800e11', 'attached_at': '', 'detached_at': '', 'volume_id': '02e72028-7ac8-4eab-88b0-9de83339ce64', 'serial': '02e72028-7ac8-4eab-88b0-9de83339ce64'} {{(pid=63531) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 943.121094] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f153b79f-dfcb-4219-a880-6506ca6cb021 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.146369] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6805d783-78e3-4910-873f-b443190ae492 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.154394] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6eeb9ff-c0b9-470f-b12f-316292d1de16 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.175870] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fff05f6d-fe97-4593-ae07-d9db71d77a25 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.193489] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-c8a73bd9-9696-4dd3-baf2-1c46189464fe tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] The volume has not been displaced from its original location: [datastore1] volume-02e72028-7ac8-4eab-88b0-9de83339ce64/volume-02e72028-7ac8-4eab-88b0-9de83339ce64.vmdk. No consolidation needed. {{(pid=63531) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 943.199248] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-c8a73bd9-9696-4dd3-baf2-1c46189464fe tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Reconfiguring VM instance instance-00000037 to detach disk 2001 {{(pid=63531) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 943.199799] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-33f453b5-e5b5-4646-8b16-cfd8a88a68aa {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.222253] env[63531]: DEBUG oslo_vmware.api [None req-c8a73bd9-9696-4dd3-baf2-1c46189464fe tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for the task: (returnval){ [ 943.222253] env[63531]: value = "task-1118245" [ 943.222253] env[63531]: _type = "Task" [ 943.222253] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.236912] env[63531]: DEBUG oslo_vmware.api [None req-c8a73bd9-9696-4dd3-baf2-1c46189464fe tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118245, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.237564] env[63531]: DEBUG oslo_concurrency.lockutils [req-a4622ad1-054e-4cc6-8594-23f46fec8a15 req-9e651580-a717-41dd-9b38-e94368c5e8b6 service nova] Releasing lock "refresh_cache-55111b87-1e9a-4877-8689-987faa72a54e" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.341994] env[63531]: DEBUG nova.compute.manager [req-8fa36b69-0355-494a-b223-fa638f2cbcb2 req-fe8d0b8c-c87c-4100-9e0f-8b9c249d04f2 service nova] [instance: 3fb0499d-8074-4e63-be9f-380730416cc4] Received event network-vif-deleted-893b680c-06c9-4135-aceb-e275af9a88d7 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 943.342154] env[63531]: INFO nova.compute.manager [req-8fa36b69-0355-494a-b223-fa638f2cbcb2 req-fe8d0b8c-c87c-4100-9e0f-8b9c249d04f2 service nova] [instance: 3fb0499d-8074-4e63-be9f-380730416cc4] Neutron deleted interface 893b680c-06c9-4135-aceb-e275af9a88d7; detaching it from the instance and deleting it from the info cache [ 943.342333] env[63531]: DEBUG nova.network.neutron [req-8fa36b69-0355-494a-b223-fa638f2cbcb2 req-fe8d0b8c-c87c-4100-9e0f-8b9c249d04f2 service nova] [instance: 3fb0499d-8074-4e63-be9f-380730416cc4] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 943.355858] env[63531]: DEBUG oslo_vmware.api [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52a0f35d-1e25-22dd-1271-4c913c072701, 'name': SearchDatastore_Task, 'duration_secs': 0.020921} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.357653] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b559143d-5d73-4fc1-a1b3-ae09ad0d3292 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.364156] env[63531]: DEBUG oslo_vmware.api [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Waiting for the task: (returnval){ [ 943.364156] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52f43d85-db6c-7f16-4b64-cb7c5d22a989" [ 943.364156] env[63531]: _type = "Task" [ 943.364156] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.375935] env[63531]: DEBUG oslo_vmware.api [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52f43d85-db6c-7f16-4b64-cb7c5d22a989, 'name': SearchDatastore_Task} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.376220] env[63531]: DEBUG oslo_concurrency.lockutils [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.376480] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 55111b87-1e9a-4877-8689-987faa72a54e/55111b87-1e9a-4877-8689-987faa72a54e.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 943.376736] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9810efb2-266c-45a2-82c9-0b921b7df470 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.383220] env[63531]: DEBUG oslo_vmware.api [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Waiting for the task: (returnval){ [ 943.383220] env[63531]: value = "task-1118246" [ 943.383220] env[63531]: _type = "Task" [ 943.383220] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.392196] env[63531]: DEBUG oslo_vmware.api [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Task: {'id': task-1118246, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.494226] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.361s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.494903] env[63531]: DEBUG nova.compute.manager [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 943.499121] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 15.931s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.510147] env[63531]: DEBUG oslo_vmware.api [None req-3ac1abf3-9cc3-459d-89d6-e0be3bbf01b0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118235, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.593927] env[63531]: DEBUG oslo_vmware.api [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118244, 'name': PowerOnVM_Task, 'duration_secs': 0.79574} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.594427] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 943.594749] env[63531]: INFO nova.compute.manager [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Took 8.70 seconds to spawn the instance on the hypervisor. [ 943.595106] env[63531]: DEBUG nova.compute.manager [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 943.595941] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34901ddb-553a-4df6-9afa-3101beccacdc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.732853] env[63531]: DEBUG oslo_vmware.api [None req-c8a73bd9-9696-4dd3-baf2-1c46189464fe tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118245, 'name': ReconfigVM_Task, 'duration_secs': 0.311724} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.732946] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-c8a73bd9-9696-4dd3-baf2-1c46189464fe tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Reconfigured VM instance instance-00000037 to detach disk 2001 {{(pid=63531) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 943.737691] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-388f3509-34c6-448d-8a85-c8a1827ef98c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.748989] env[63531]: DEBUG nova.network.neutron [-] [instance: 3fb0499d-8074-4e63-be9f-380730416cc4] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 943.753417] env[63531]: DEBUG oslo_vmware.api [None req-c8a73bd9-9696-4dd3-baf2-1c46189464fe tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for the task: (returnval){ [ 943.753417] env[63531]: value = "task-1118247" [ 943.753417] env[63531]: _type = "Task" [ 943.753417] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.761557] env[63531]: DEBUG oslo_vmware.api [None req-c8a73bd9-9696-4dd3-baf2-1c46189464fe tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118247, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.845939] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b1451dcb-88c3-4c8f-96ed-144a6f74dcb6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.855168] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d7958f1-a95f-4c82-be06-aaf306c01c92 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.888332] env[63531]: DEBUG nova.compute.manager [req-8fa36b69-0355-494a-b223-fa638f2cbcb2 req-fe8d0b8c-c87c-4100-9e0f-8b9c249d04f2 service nova] [instance: 3fb0499d-8074-4e63-be9f-380730416cc4] Detach interface failed, port_id=893b680c-06c9-4135-aceb-e275af9a88d7, reason: Instance 3fb0499d-8074-4e63-be9f-380730416cc4 could not be found. {{(pid=63531) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 943.896851] env[63531]: DEBUG oslo_vmware.api [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Task: {'id': task-1118246, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.022100] env[63531]: DEBUG nova.compute.utils [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 944.024634] env[63531]: DEBUG nova.compute.manager [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 944.024993] env[63531]: DEBUG nova.network.neutron [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 944.038063] env[63531]: DEBUG oslo_vmware.api [None req-3ac1abf3-9cc3-459d-89d6-e0be3bbf01b0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118235, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.101390] env[63531]: DEBUG nova.policy [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0a501b2c64c04dcda3596c8a1bd3fa0d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'abdae960d43b47bc95ab41f265d2c85b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 944.120671] env[63531]: INFO nova.compute.manager [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Took 32.37 seconds to build instance. [ 944.251570] env[63531]: INFO nova.compute.manager [-] [instance: 3fb0499d-8074-4e63-be9f-380730416cc4] Took 1.24 seconds to deallocate network for instance. [ 944.264864] env[63531]: DEBUG oslo_vmware.api [None req-c8a73bd9-9696-4dd3-baf2-1c46189464fe tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118247, 'name': ReconfigVM_Task, 'duration_secs': 0.178484} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.265169] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-c8a73bd9-9696-4dd3-baf2-1c46189464fe tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244696', 'volume_id': '02e72028-7ac8-4eab-88b0-9de83339ce64', 'name': 'volume-02e72028-7ac8-4eab-88b0-9de83339ce64', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '48043b5d-c0d3-4e8a-8d28-5a408d800e11', 'attached_at': '', 'detached_at': '', 'volume_id': '02e72028-7ac8-4eab-88b0-9de83339ce64', 'serial': '02e72028-7ac8-4eab-88b0-9de83339ce64'} {{(pid=63531) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 944.398018] env[63531]: DEBUG oslo_vmware.api [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Task: {'id': task-1118246, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.857602} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.398340] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 55111b87-1e9a-4877-8689-987faa72a54e/55111b87-1e9a-4877-8689-987faa72a54e.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 944.398562] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] [instance: 55111b87-1e9a-4877-8689-987faa72a54e] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 944.398828] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e7d89b1e-a404-4af5-a01e-d2eea28e7580 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.404907] env[63531]: DEBUG oslo_vmware.api [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Waiting for the task: (returnval){ [ 944.404907] env[63531]: value = "task-1118248" [ 944.404907] env[63531]: _type = "Task" [ 944.404907] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.412429] env[63531]: DEBUG oslo_vmware.api [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Task: {'id': task-1118248, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.504626] env[63531]: DEBUG nova.network.neutron [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Successfully created port: d3f47a79-f699-4d63-9412-724fe1f15913 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 944.512099] env[63531]: DEBUG oslo_vmware.api [None req-3ac1abf3-9cc3-459d-89d6-e0be3bbf01b0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118235, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.526113] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Applying migration context for instance 9a3fad46-a15a-451c-bdab-a3c8cc8add07 as it has an incoming, in-progress migration 72374b80-67c2-4204-8fdb-7c6eff83d384. Migration status is confirming {{(pid=63531) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 944.528017] env[63531]: INFO nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Updating resource usage from migration 72374b80-67c2-4204-8fdb-7c6eff83d384 [ 944.531737] env[63531]: DEBUG nova.compute.manager [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 944.554033] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 29dd6de3-2765-408c-acc0-da47e5e0a977 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 944.554212] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 8dec0cac-defc-42f2-ab56-6b3ae60ad858 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 944.554341] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 48043b5d-c0d3-4e8a-8d28-5a408d800e11 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 944.554462] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance f46d8a85-6ab7-436a-bc5e-40d0f80974e5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 944.554578] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 944.554692] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 879924d4-b465-4102-a0e3-c7b2be7ef08b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 944.554814] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance a090135f-489d-457d-be36-ba6f61b71ab8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 944.554928] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 944.555051] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 944.555193] env[63531]: WARNING nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 6118e242-c0db-4707-a5f9-3d12b823935f is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 944.555318] env[63531]: WARNING nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 4b5e5821-9def-4b53-be19-355d9e5f81ec is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 944.555433] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Migration 72374b80-67c2-4204-8fdb-7c6eff83d384 is active on this compute host and has allocations in placement: {'resources': {'VCPU': 1, 'MEMORY_MB': 192, 'DISK_GB': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 944.555604] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 9a3fad46-a15a-451c-bdab-a3c8cc8add07 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 944.555695] env[63531]: WARNING nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance e065f052-ba3e-4783-8953-5dc200d1f3e9 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 944.555774] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance ccf00b09-29fe-4a76-a8af-97774acd77de actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 944.555883] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 3fb0499d-8074-4e63-be9f-380730416cc4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 944.622713] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5d7ee46e-3908-4b7a-b152-64181c52582b tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "27c23b7c-a68f-43c9-a125-bd657feb3c5b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.879s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.761245] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ba2842a1-0c0c-4cb4-b87b-242607074b48 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.811075] env[63531]: DEBUG nova.objects.instance [None req-c8a73bd9-9696-4dd3-baf2-1c46189464fe tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lazy-loading 'flavor' on Instance uuid 48043b5d-c0d3-4e8a-8d28-5a408d800e11 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 944.915272] env[63531]: DEBUG oslo_vmware.api [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Task: {'id': task-1118248, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082558} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.915626] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] [instance: 55111b87-1e9a-4877-8689-987faa72a54e] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 944.916449] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-637e5823-7e04-4b4e-8bea-45f8d7f4ce97 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.939242] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] [instance: 55111b87-1e9a-4877-8689-987faa72a54e] Reconfiguring VM instance instance-00000054 to attach disk [datastore1] 55111b87-1e9a-4877-8689-987faa72a54e/55111b87-1e9a-4877-8689-987faa72a54e.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 944.939579] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c8de60dd-501f-4a84-9bb6-351a7f1456c2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.961537] env[63531]: DEBUG oslo_vmware.api [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Waiting for the task: (returnval){ [ 944.961537] env[63531]: value = "task-1118249" [ 944.961537] env[63531]: _type = "Task" [ 944.961537] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.968933] env[63531]: DEBUG oslo_vmware.api [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Task: {'id': task-1118249, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.009901] env[63531]: DEBUG oslo_vmware.api [None req-3ac1abf3-9cc3-459d-89d6-e0be3bbf01b0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118235, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.061023] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 1ca7b851-2a06-4181-8271-58aafcd322d6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 945.061023] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 27c23b7c-a68f-43c9-a125-bd657feb3c5b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 945.061023] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 55111b87-1e9a-4877-8689-987faa72a54e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 945.061023] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 911f23f7-e320-48f3-87a9-0239b013feff actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 945.470833] env[63531]: DEBUG oslo_vmware.api [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Task: {'id': task-1118249, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.511023] env[63531]: DEBUG oslo_vmware.api [None req-3ac1abf3-9cc3-459d-89d6-e0be3bbf01b0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118235, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.548042] env[63531]: DEBUG nova.compute.manager [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 945.565773] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 945.565773] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Total usable vcpus: 48, total allocated vcpus: 16 {{(pid=63531) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 945.565773] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3648MB phys_disk=200GB used_disk=16GB total_vcpus=48 used_vcpus=16 pci_stats=[] {{(pid=63531) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 945.577645] env[63531]: DEBUG nova.virt.hardware [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 945.577986] env[63531]: DEBUG nova.virt.hardware [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 945.578180] env[63531]: DEBUG nova.virt.hardware [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 945.578372] env[63531]: DEBUG nova.virt.hardware [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 945.578522] env[63531]: DEBUG nova.virt.hardware [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 945.578673] env[63531]: DEBUG nova.virt.hardware [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 945.578886] env[63531]: DEBUG nova.virt.hardware [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 945.579090] env[63531]: DEBUG nova.virt.hardware [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 945.579284] env[63531]: DEBUG nova.virt.hardware [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 945.579456] env[63531]: DEBUG nova.virt.hardware [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 945.579635] env[63531]: DEBUG nova.virt.hardware [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 945.580561] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08989dd7-5494-45af-94b7-9dbb5d3b95b3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.592324] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53e1b195-2ca3-4119-8d9b-09936bbb563f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.789205] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5bb1734-11fe-4310-a6aa-df6681f3619c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.796864] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48b93583-b941-43d2-b9e6-b16e49002dc0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.827180] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c8a73bd9-9696-4dd3-baf2-1c46189464fe tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "48043b5d-c0d3-4e8a-8d28-5a408d800e11" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.249s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.829055] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b5e40e3-700b-493f-aa1a-1832e08c67ef {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.837015] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82442d20-7986-4036-9248-27e4c65524be {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.854446] env[63531]: DEBUG nova.compute.provider_tree [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 945.972173] env[63531]: DEBUG oslo_vmware.api [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Task: {'id': task-1118249, 'name': ReconfigVM_Task, 'duration_secs': 0.933793} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.972504] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] [instance: 55111b87-1e9a-4877-8689-987faa72a54e] Reconfigured VM instance instance-00000054 to attach disk [datastore1] 55111b87-1e9a-4877-8689-987faa72a54e/55111b87-1e9a-4877-8689-987faa72a54e.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 945.973204] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6a8c958e-2b2b-41c2-a97e-c4b63070256f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.981031] env[63531]: DEBUG oslo_vmware.api [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Waiting for the task: (returnval){ [ 945.981031] env[63531]: value = "task-1118250" [ 945.981031] env[63531]: _type = "Task" [ 945.981031] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.988886] env[63531]: DEBUG oslo_vmware.api [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Task: {'id': task-1118250, 'name': Rename_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.012324] env[63531]: DEBUG oslo_vmware.api [None req-3ac1abf3-9cc3-459d-89d6-e0be3bbf01b0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118235, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.078638] env[63531]: DEBUG nova.compute.manager [req-bfefaa3b-b108-45b5-bfa4-6af881b991ab req-74a08442-4632-4412-9946-d5ce817e8d52 service nova] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Received event network-vif-plugged-d3f47a79-f699-4d63-9412-724fe1f15913 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 946.078907] env[63531]: DEBUG oslo_concurrency.lockutils [req-bfefaa3b-b108-45b5-bfa4-6af881b991ab req-74a08442-4632-4412-9946-d5ce817e8d52 service nova] Acquiring lock "911f23f7-e320-48f3-87a9-0239b013feff-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.079163] env[63531]: DEBUG oslo_concurrency.lockutils [req-bfefaa3b-b108-45b5-bfa4-6af881b991ab req-74a08442-4632-4412-9946-d5ce817e8d52 service nova] Lock "911f23f7-e320-48f3-87a9-0239b013feff-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.079671] env[63531]: DEBUG oslo_concurrency.lockutils [req-bfefaa3b-b108-45b5-bfa4-6af881b991ab req-74a08442-4632-4412-9946-d5ce817e8d52 service nova] Lock "911f23f7-e320-48f3-87a9-0239b013feff-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.079938] env[63531]: DEBUG nova.compute.manager [req-bfefaa3b-b108-45b5-bfa4-6af881b991ab req-74a08442-4632-4412-9946-d5ce817e8d52 service nova] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] No waiting events found dispatching network-vif-plugged-d3f47a79-f699-4d63-9412-724fe1f15913 {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 946.080215] env[63531]: WARNING nova.compute.manager [req-bfefaa3b-b108-45b5-bfa4-6af881b991ab req-74a08442-4632-4412-9946-d5ce817e8d52 service nova] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Received unexpected event network-vif-plugged-d3f47a79-f699-4d63-9412-724fe1f15913 for instance with vm_state building and task_state spawning. [ 946.241573] env[63531]: DEBUG nova.network.neutron [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Successfully updated port: d3f47a79-f699-4d63-9412-724fe1f15913 {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 946.360019] env[63531]: DEBUG nova.scheduler.client.report [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 946.490926] env[63531]: DEBUG oslo_vmware.api [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Task: {'id': task-1118250, 'name': Rename_Task, 'duration_secs': 0.1636} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.491496] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] [instance: 55111b87-1e9a-4877-8689-987faa72a54e] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 946.491888] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-952814c9-371e-4d84-8094-a5d691b3ca2e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.498503] env[63531]: DEBUG oslo_vmware.api [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Waiting for the task: (returnval){ [ 946.498503] env[63531]: value = "task-1118251" [ 946.498503] env[63531]: _type = "Task" [ 946.498503] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.509322] env[63531]: DEBUG oslo_vmware.api [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Task: {'id': task-1118251, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.515432] env[63531]: DEBUG oslo_vmware.api [None req-3ac1abf3-9cc3-459d-89d6-e0be3bbf01b0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118235, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.659578] env[63531]: DEBUG nova.compute.manager [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Stashing vm_state: active {{(pid=63531) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 946.744580] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquiring lock "refresh_cache-911f23f7-e320-48f3-87a9-0239b013feff" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 946.744580] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquired lock "refresh_cache-911f23f7-e320-48f3-87a9-0239b013feff" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.744580] env[63531]: DEBUG nova.network.neutron [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 946.862911] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63531) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 946.863139] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.364s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.863432] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 18.599s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.863619] env[63531]: DEBUG nova.objects.instance [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63531) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 947.009563] env[63531]: DEBUG oslo_vmware.api [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Task: {'id': task-1118251, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.014776] env[63531]: DEBUG oslo_vmware.api [None req-3ac1abf3-9cc3-459d-89d6-e0be3bbf01b0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118235, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.019335] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a4bb48b8-dfee-4f65-99b6-ae50dabf2571 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquiring lock "48043b5d-c0d3-4e8a-8d28-5a408d800e11" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.019607] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a4bb48b8-dfee-4f65-99b6-ae50dabf2571 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "48043b5d-c0d3-4e8a-8d28-5a408d800e11" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.019853] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a4bb48b8-dfee-4f65-99b6-ae50dabf2571 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquiring lock "48043b5d-c0d3-4e8a-8d28-5a408d800e11-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.020132] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a4bb48b8-dfee-4f65-99b6-ae50dabf2571 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "48043b5d-c0d3-4e8a-8d28-5a408d800e11-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.020347] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a4bb48b8-dfee-4f65-99b6-ae50dabf2571 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "48043b5d-c0d3-4e8a-8d28-5a408d800e11-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.022580] env[63531]: INFO nova.compute.manager [None req-a4bb48b8-dfee-4f65-99b6-ae50dabf2571 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Terminating instance [ 947.024527] env[63531]: DEBUG nova.compute.manager [None req-a4bb48b8-dfee-4f65-99b6-ae50dabf2571 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 947.024724] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a4bb48b8-dfee-4f65-99b6-ae50dabf2571 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 947.025734] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ecea00c-00ea-424b-95a6-7988031ee34f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.032714] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4bb48b8-dfee-4f65-99b6-ae50dabf2571 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 947.032935] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b6c86d1a-7b38-4433-b069-1362b7d2f7a0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.039368] env[63531]: DEBUG oslo_vmware.api [None req-a4bb48b8-dfee-4f65-99b6-ae50dabf2571 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for the task: (returnval){ [ 947.039368] env[63531]: value = "task-1118252" [ 947.039368] env[63531]: _type = "Task" [ 947.039368] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.046550] env[63531]: DEBUG oslo_vmware.api [None req-a4bb48b8-dfee-4f65-99b6-ae50dabf2571 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118252, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.178109] env[63531]: DEBUG oslo_concurrency.lockutils [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.281216] env[63531]: DEBUG nova.network.neutron [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 947.514527] env[63531]: DEBUG oslo_vmware.api [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Task: {'id': task-1118251, 'name': PowerOnVM_Task, 'duration_secs': 0.653892} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.517642] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] [instance: 55111b87-1e9a-4877-8689-987faa72a54e] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 947.518250] env[63531]: INFO nova.compute.manager [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] [instance: 55111b87-1e9a-4877-8689-987faa72a54e] Took 8.23 seconds to spawn the instance on the hypervisor. [ 947.518573] env[63531]: DEBUG nova.compute.manager [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] [instance: 55111b87-1e9a-4877-8689-987faa72a54e] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 947.519102] env[63531]: DEBUG oslo_vmware.api [None req-3ac1abf3-9cc3-459d-89d6-e0be3bbf01b0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118235, 'name': ReconfigVM_Task, 'duration_secs': 6.338641} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.520039] env[63531]: DEBUG nova.network.neutron [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Updating instance_info_cache with network_info: [{"id": "d3f47a79-f699-4d63-9412-724fe1f15913", "address": "fa:16:3e:71:30:b9", "network": {"id": "eac8786d-bc67-4916-8483-b53d9a25efb8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1604989690-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "abdae960d43b47bc95ab41f265d2c85b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db068f71-08cc-42d4-8ab6-17134c1585e5", "external-id": "nsx-vlan-transportzone-721", "segmentation_id": 721, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd3f47a79-f6", "ovs_interfaceid": "d3f47a79-f699-4d63-9412-724fe1f15913", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 947.521771] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18cfb52f-a501-4c60-a14e-b35871abbd39 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.524248] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3ac1abf3-9cc3-459d-89d6-e0be3bbf01b0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Releasing lock "a090135f-489d-457d-be36-ba6f61b71ab8" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.524592] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-3ac1abf3-9cc3-459d-89d6-e0be3bbf01b0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Reconfigured VM to detach interface {{(pid=63531) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 947.548136] env[63531]: DEBUG oslo_vmware.api [None req-a4bb48b8-dfee-4f65-99b6-ae50dabf2571 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118252, 'name': PowerOffVM_Task, 'duration_secs': 0.260522} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.548630] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4bb48b8-dfee-4f65-99b6-ae50dabf2571 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 947.548954] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a4bb48b8-dfee-4f65-99b6-ae50dabf2571 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 947.549315] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-74fc9d21-0aa9-4c39-bfd6-99dc5f5f3a08 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.640024] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a4bb48b8-dfee-4f65-99b6-ae50dabf2571 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 947.640024] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a4bb48b8-dfee-4f65-99b6-ae50dabf2571 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Deleting contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 947.640024] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4bb48b8-dfee-4f65-99b6-ae50dabf2571 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Deleting the datastore file [datastore2] 48043b5d-c0d3-4e8a-8d28-5a408d800e11 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 947.640024] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7808e2c6-ce70-418a-9384-2ea804d20ece {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.644803] env[63531]: DEBUG oslo_vmware.api [None req-a4bb48b8-dfee-4f65-99b6-ae50dabf2571 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for the task: (returnval){ [ 947.644803] env[63531]: value = "task-1118254" [ 947.644803] env[63531]: _type = "Task" [ 947.644803] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.653209] env[63531]: DEBUG oslo_vmware.api [None req-a4bb48b8-dfee-4f65-99b6-ae50dabf2571 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118254, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.875029] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c1592367-cf54-40d6-a3bc-3415c963e362 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.009s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.875029] env[63531]: DEBUG oslo_concurrency.lockutils [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.849s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.875029] env[63531]: DEBUG nova.objects.instance [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Lazy-loading 'pci_requests' on Instance uuid 1ca7b851-2a06-4181-8271-58aafcd322d6 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 948.027586] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Releasing lock "refresh_cache-911f23f7-e320-48f3-87a9-0239b013feff" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 948.027996] env[63531]: DEBUG nova.compute.manager [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Instance network_info: |[{"id": "d3f47a79-f699-4d63-9412-724fe1f15913", "address": "fa:16:3e:71:30:b9", "network": {"id": "eac8786d-bc67-4916-8483-b53d9a25efb8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1604989690-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "abdae960d43b47bc95ab41f265d2c85b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db068f71-08cc-42d4-8ab6-17134c1585e5", "external-id": "nsx-vlan-transportzone-721", "segmentation_id": 721, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd3f47a79-f6", "ovs_interfaceid": "d3f47a79-f699-4d63-9412-724fe1f15913", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 948.028478] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:71:30:b9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'db068f71-08cc-42d4-8ab6-17134c1585e5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd3f47a79-f699-4d63-9412-724fe1f15913', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 948.036288] env[63531]: DEBUG oslo.service.loopingcall [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 948.040447] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 948.042638] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1547f01d-3aa9-4e46-9f70-f13c91c0dfe7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.056924] env[63531]: INFO nova.compute.manager [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] [instance: 55111b87-1e9a-4877-8689-987faa72a54e] Took 34.53 seconds to build instance. [ 948.063589] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 948.063589] env[63531]: value = "task-1118255" [ 948.063589] env[63531]: _type = "Task" [ 948.063589] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.071754] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118255, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.111748] env[63531]: DEBUG nova.compute.manager [req-2dbf38a6-5e61-4daf-afb5-6104396969e1 req-bc7e5bb5-df3f-4730-94ed-e4e09f0e4e8c service nova] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Received event network-changed-d3f47a79-f699-4d63-9412-724fe1f15913 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 948.112195] env[63531]: DEBUG nova.compute.manager [req-2dbf38a6-5e61-4daf-afb5-6104396969e1 req-bc7e5bb5-df3f-4730-94ed-e4e09f0e4e8c service nova] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Refreshing instance network info cache due to event network-changed-d3f47a79-f699-4d63-9412-724fe1f15913. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 948.112518] env[63531]: DEBUG oslo_concurrency.lockutils [req-2dbf38a6-5e61-4daf-afb5-6104396969e1 req-bc7e5bb5-df3f-4730-94ed-e4e09f0e4e8c service nova] Acquiring lock "refresh_cache-911f23f7-e320-48f3-87a9-0239b013feff" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.112789] env[63531]: DEBUG oslo_concurrency.lockutils [req-2dbf38a6-5e61-4daf-afb5-6104396969e1 req-bc7e5bb5-df3f-4730-94ed-e4e09f0e4e8c service nova] Acquired lock "refresh_cache-911f23f7-e320-48f3-87a9-0239b013feff" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.113008] env[63531]: DEBUG nova.network.neutron [req-2dbf38a6-5e61-4daf-afb5-6104396969e1 req-bc7e5bb5-df3f-4730-94ed-e4e09f0e4e8c service nova] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Refreshing network info cache for port d3f47a79-f699-4d63-9412-724fe1f15913 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 948.157176] env[63531]: DEBUG oslo_vmware.api [None req-a4bb48b8-dfee-4f65-99b6-ae50dabf2571 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118254, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.427212} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.157402] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4bb48b8-dfee-4f65-99b6-ae50dabf2571 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 948.157596] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a4bb48b8-dfee-4f65-99b6-ae50dabf2571 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Deleted contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 948.157826] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a4bb48b8-dfee-4f65-99b6-ae50dabf2571 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 948.158022] env[63531]: INFO nova.compute.manager [None req-a4bb48b8-dfee-4f65-99b6-ae50dabf2571 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Took 1.13 seconds to destroy the instance on the hypervisor. [ 948.158271] env[63531]: DEBUG oslo.service.loopingcall [None req-a4bb48b8-dfee-4f65-99b6-ae50dabf2571 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 948.158515] env[63531]: DEBUG nova.compute.manager [-] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 948.158638] env[63531]: DEBUG nova.network.neutron [-] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 948.378085] env[63531]: DEBUG nova.objects.instance [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Lazy-loading 'numa_topology' on Instance uuid 1ca7b851-2a06-4181-8271-58aafcd322d6 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 948.559324] env[63531]: DEBUG oslo_concurrency.lockutils [None req-813a9712-5bee-49d7-b182-8da73ee87f91 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Lock "55111b87-1e9a-4877-8689-987faa72a54e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.042s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.573827] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118255, 'name': CreateVM_Task} progress is 25%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.851134] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3ac1abf3-9cc3-459d-89d6-e0be3bbf01b0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "refresh_cache-a090135f-489d-457d-be36-ba6f61b71ab8" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.853104] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3ac1abf3-9cc3-459d-89d6-e0be3bbf01b0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquired lock "refresh_cache-a090135f-489d-457d-be36-ba6f61b71ab8" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.853104] env[63531]: DEBUG nova.network.neutron [None req-3ac1abf3-9cc3-459d-89d6-e0be3bbf01b0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 948.854618] env[63531]: DEBUG nova.network.neutron [req-2dbf38a6-5e61-4daf-afb5-6104396969e1 req-bc7e5bb5-df3f-4730-94ed-e4e09f0e4e8c service nova] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Updated VIF entry in instance network info cache for port d3f47a79-f699-4d63-9412-724fe1f15913. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 948.855174] env[63531]: DEBUG nova.network.neutron [req-2dbf38a6-5e61-4daf-afb5-6104396969e1 req-bc7e5bb5-df3f-4730-94ed-e4e09f0e4e8c service nova] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Updating instance_info_cache with network_info: [{"id": "d3f47a79-f699-4d63-9412-724fe1f15913", "address": "fa:16:3e:71:30:b9", "network": {"id": "eac8786d-bc67-4916-8483-b53d9a25efb8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1604989690-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "abdae960d43b47bc95ab41f265d2c85b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db068f71-08cc-42d4-8ab6-17134c1585e5", "external-id": "nsx-vlan-transportzone-721", "segmentation_id": 721, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd3f47a79-f6", "ovs_interfaceid": "d3f47a79-f699-4d63-9412-724fe1f15913", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 948.880419] env[63531]: INFO nova.compute.claims [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 949.077090] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118255, 'name': CreateVM_Task, 'duration_secs': 0.854926} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.077090] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 949.077700] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 949.077908] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.079113] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 949.079113] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-12ce42d1-9a2a-4f22-a492-a6ddb66c59d6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.086234] env[63531]: DEBUG oslo_vmware.api [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 949.086234] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52d8ae58-57a5-1214-bc55-d992d0aeda0b" [ 949.086234] env[63531]: _type = "Task" [ 949.086234] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.094383] env[63531]: DEBUG oslo_vmware.api [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52d8ae58-57a5-1214-bc55-d992d0aeda0b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.360182] env[63531]: DEBUG oslo_concurrency.lockutils [req-2dbf38a6-5e61-4daf-afb5-6104396969e1 req-bc7e5bb5-df3f-4730-94ed-e4e09f0e4e8c service nova] Releasing lock "refresh_cache-911f23f7-e320-48f3-87a9-0239b013feff" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 949.360460] env[63531]: DEBUG nova.compute.manager [req-2dbf38a6-5e61-4daf-afb5-6104396969e1 req-bc7e5bb5-df3f-4730-94ed-e4e09f0e4e8c service nova] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Received event network-vif-deleted-1d5fea6d-10c4-4740-81a1-8e1ac927dad5 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 949.360644] env[63531]: INFO nova.compute.manager [req-2dbf38a6-5e61-4daf-afb5-6104396969e1 req-bc7e5bb5-df3f-4730-94ed-e4e09f0e4e8c service nova] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Neutron deleted interface 1d5fea6d-10c4-4740-81a1-8e1ac927dad5; detaching it from the instance and deleting it from the info cache [ 949.360902] env[63531]: DEBUG nova.network.neutron [req-2dbf38a6-5e61-4daf-afb5-6104396969e1 req-bc7e5bb5-df3f-4730-94ed-e4e09f0e4e8c service nova] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Updating instance_info_cache with network_info: [{"id": "17da826c-920c-4044-93cb-be8c8cf01c13", "address": "fa:16:3e:9a:ac:ad", "network": {"id": "caf86428-59a5-4e50-9135-2ab8473af64d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1673994696-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.214", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ba8b284854242c392aec5326e996239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap17da826c-92", "ovs_interfaceid": "17da826c-920c-4044-93cb-be8c8cf01c13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 949.494727] env[63531]: DEBUG nova.network.neutron [-] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 949.567562] env[63531]: INFO nova.network.neutron [None req-3ac1abf3-9cc3-459d-89d6-e0be3bbf01b0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Port 1d5fea6d-10c4-4740-81a1-8e1ac927dad5 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 949.568171] env[63531]: DEBUG nova.network.neutron [None req-3ac1abf3-9cc3-459d-89d6-e0be3bbf01b0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Updating instance_info_cache with network_info: [{"id": "17da826c-920c-4044-93cb-be8c8cf01c13", "address": "fa:16:3e:9a:ac:ad", "network": {"id": "caf86428-59a5-4e50-9135-2ab8473af64d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1673994696-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.214", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ba8b284854242c392aec5326e996239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap17da826c-92", "ovs_interfaceid": "17da826c-920c-4044-93cb-be8c8cf01c13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 949.598868] env[63531]: DEBUG oslo_vmware.api [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52d8ae58-57a5-1214-bc55-d992d0aeda0b, 'name': SearchDatastore_Task, 'duration_secs': 0.010948} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.599179] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 949.599463] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 949.599624] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 949.599831] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.600096] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 949.600379] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2dbe1066-508b-4608-89d8-052c484cc491 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.608731] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 949.608917] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 949.609631] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0d265bab-4ba9-4606-a5e6-50dd9150f9ec {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.614496] env[63531]: DEBUG oslo_vmware.api [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 949.614496] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]526ba309-1483-7b08-0184-19119b43d1d5" [ 949.614496] env[63531]: _type = "Task" [ 949.614496] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.622608] env[63531]: DEBUG oslo_vmware.api [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]526ba309-1483-7b08-0184-19119b43d1d5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.640344] env[63531]: DEBUG oslo_concurrency.lockutils [None req-04a987cd-eb1f-407a-8135-fd56b4798069 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Acquiring lock "55111b87-1e9a-4877-8689-987faa72a54e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.640502] env[63531]: DEBUG oslo_concurrency.lockutils [None req-04a987cd-eb1f-407a-8135-fd56b4798069 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Lock "55111b87-1e9a-4877-8689-987faa72a54e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.640713] env[63531]: DEBUG oslo_concurrency.lockutils [None req-04a987cd-eb1f-407a-8135-fd56b4798069 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Acquiring lock "55111b87-1e9a-4877-8689-987faa72a54e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.640954] env[63531]: DEBUG oslo_concurrency.lockutils [None req-04a987cd-eb1f-407a-8135-fd56b4798069 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Lock "55111b87-1e9a-4877-8689-987faa72a54e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.641197] env[63531]: DEBUG oslo_concurrency.lockutils [None req-04a987cd-eb1f-407a-8135-fd56b4798069 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Lock "55111b87-1e9a-4877-8689-987faa72a54e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.643007] env[63531]: DEBUG oslo_concurrency.lockutils [None req-293635f3-1f0c-487a-ad12-e4a71b80c8dd tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "a090135f-489d-457d-be36-ba6f61b71ab8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.643277] env[63531]: DEBUG oslo_concurrency.lockutils [None req-293635f3-1f0c-487a-ad12-e4a71b80c8dd tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "a090135f-489d-457d-be36-ba6f61b71ab8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.643500] env[63531]: DEBUG oslo_concurrency.lockutils [None req-293635f3-1f0c-487a-ad12-e4a71b80c8dd tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "a090135f-489d-457d-be36-ba6f61b71ab8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.643721] env[63531]: DEBUG oslo_concurrency.lockutils [None req-293635f3-1f0c-487a-ad12-e4a71b80c8dd tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "a090135f-489d-457d-be36-ba6f61b71ab8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.643919] env[63531]: DEBUG oslo_concurrency.lockutils [None req-293635f3-1f0c-487a-ad12-e4a71b80c8dd tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "a090135f-489d-457d-be36-ba6f61b71ab8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.645542] env[63531]: INFO nova.compute.manager [None req-04a987cd-eb1f-407a-8135-fd56b4798069 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] [instance: 55111b87-1e9a-4877-8689-987faa72a54e] Terminating instance [ 949.646892] env[63531]: INFO nova.compute.manager [None req-293635f3-1f0c-487a-ad12-e4a71b80c8dd tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Terminating instance [ 949.648658] env[63531]: DEBUG nova.compute.manager [None req-293635f3-1f0c-487a-ad12-e4a71b80c8dd tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 949.648849] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-293635f3-1f0c-487a-ad12-e4a71b80c8dd tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 949.649387] env[63531]: DEBUG nova.compute.manager [None req-04a987cd-eb1f-407a-8135-fd56b4798069 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] [instance: 55111b87-1e9a-4877-8689-987faa72a54e] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 949.649568] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-04a987cd-eb1f-407a-8135-fd56b4798069 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] [instance: 55111b87-1e9a-4877-8689-987faa72a54e] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 949.650401] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08d83baf-a93b-4b14-b500-e30e720362c2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.653844] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b89b9d5d-c7a2-41e9-a8cd-be405c579f62 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.661110] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-04a987cd-eb1f-407a-8135-fd56b4798069 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] [instance: 55111b87-1e9a-4877-8689-987faa72a54e] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 949.663329] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e6fca825-292d-48f1-98d3-0768f08179e2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.664502] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-293635f3-1f0c-487a-ad12-e4a71b80c8dd tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 949.664718] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8c3175e4-5d25-4f13-acf8-c7fa3000bb5b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.670447] env[63531]: DEBUG oslo_vmware.api [None req-293635f3-1f0c-487a-ad12-e4a71b80c8dd tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for the task: (returnval){ [ 949.670447] env[63531]: value = "task-1118257" [ 949.670447] env[63531]: _type = "Task" [ 949.670447] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.671552] env[63531]: DEBUG oslo_vmware.api [None req-04a987cd-eb1f-407a-8135-fd56b4798069 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Waiting for the task: (returnval){ [ 949.671552] env[63531]: value = "task-1118256" [ 949.671552] env[63531]: _type = "Task" [ 949.671552] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.681879] env[63531]: DEBUG oslo_vmware.api [None req-293635f3-1f0c-487a-ad12-e4a71b80c8dd tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118257, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.684647] env[63531]: DEBUG oslo_vmware.api [None req-04a987cd-eb1f-407a-8135-fd56b4798069 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Task: {'id': task-1118256, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.867299] env[63531]: DEBUG oslo_concurrency.lockutils [req-2dbf38a6-5e61-4daf-afb5-6104396969e1 req-bc7e5bb5-df3f-4730-94ed-e4e09f0e4e8c service nova] Acquiring lock "a090135f-489d-457d-be36-ba6f61b71ab8" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 949.999347] env[63531]: INFO nova.compute.manager [-] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Took 1.84 seconds to deallocate network for instance. [ 950.071019] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3ac1abf3-9cc3-459d-89d6-e0be3bbf01b0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Releasing lock "refresh_cache-a090135f-489d-457d-be36-ba6f61b71ab8" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 950.126829] env[63531]: DEBUG oslo_vmware.api [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]526ba309-1483-7b08-0184-19119b43d1d5, 'name': SearchDatastore_Task, 'duration_secs': 0.00792} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.127635] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-08d7bd21-4e57-46bb-b3bb-d2f8197b5f0f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.136039] env[63531]: DEBUG oslo_vmware.api [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 950.136039] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]526a9722-1822-fba7-c11a-1e853a1b1a94" [ 950.136039] env[63531]: _type = "Task" [ 950.136039] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.144041] env[63531]: DEBUG oslo_vmware.api [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]526a9722-1822-fba7-c11a-1e853a1b1a94, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.167841] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb4cf764-7e89-412b-a852-d6330439c676 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.172050] env[63531]: DEBUG nova.compute.manager [req-b76378dc-fdbf-4a45-956c-3adc5a71941b req-275b7a47-ee8a-4ca0-a73b-332a4a64f8ae service nova] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Received event network-vif-deleted-44f13e67-ffdf-401f-bdfe-fbd57e4f4d63 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 950.181426] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5ff6c50-ea04-4a86-b9d6-ed46affe3a57 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.187663] env[63531]: DEBUG oslo_vmware.api [None req-293635f3-1f0c-487a-ad12-e4a71b80c8dd tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118257, 'name': PowerOffVM_Task, 'duration_secs': 0.25185} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.188221] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-293635f3-1f0c-487a-ad12-e4a71b80c8dd tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 950.188421] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-293635f3-1f0c-487a-ad12-e4a71b80c8dd tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 950.188655] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-14b9ac67-de09-4414-9fc0-687d5c18fa38 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.218476] env[63531]: DEBUG oslo_vmware.api [None req-04a987cd-eb1f-407a-8135-fd56b4798069 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Task: {'id': task-1118256, 'name': PowerOffVM_Task, 'duration_secs': 0.251746} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.219844] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-04a987cd-eb1f-407a-8135-fd56b4798069 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] [instance: 55111b87-1e9a-4877-8689-987faa72a54e] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 950.220049] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-04a987cd-eb1f-407a-8135-fd56b4798069 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] [instance: 55111b87-1e9a-4877-8689-987faa72a54e] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 950.220771] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f728489-1428-4849-b014-60505e4d1727 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.223177] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f1e896ca-ef84-4234-a623-aab13ae941c9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.229041] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa494fa6-f764-4f0a-ad49-11863dfeced5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.242150] env[63531]: DEBUG nova.compute.provider_tree [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 950.316900] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-293635f3-1f0c-487a-ad12-e4a71b80c8dd tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 950.317157] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-293635f3-1f0c-487a-ad12-e4a71b80c8dd tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 950.317341] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-293635f3-1f0c-487a-ad12-e4a71b80c8dd tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Deleting the datastore file [datastore1] a090135f-489d-457d-be36-ba6f61b71ab8 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 950.317607] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4fde61d6-055b-45d2-bc41-f22786cff2e1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.325460] env[63531]: DEBUG oslo_vmware.api [None req-293635f3-1f0c-487a-ad12-e4a71b80c8dd tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for the task: (returnval){ [ 950.325460] env[63531]: value = "task-1118260" [ 950.325460] env[63531]: _type = "Task" [ 950.325460] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.329569] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-04a987cd-eb1f-407a-8135-fd56b4798069 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] [instance: 55111b87-1e9a-4877-8689-987faa72a54e] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 950.329778] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-04a987cd-eb1f-407a-8135-fd56b4798069 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] [instance: 55111b87-1e9a-4877-8689-987faa72a54e] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 950.329968] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-04a987cd-eb1f-407a-8135-fd56b4798069 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Deleting the datastore file [datastore1] 55111b87-1e9a-4877-8689-987faa72a54e {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 950.330742] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3e0ced1a-43b2-4dcb-a98e-14d30aa4d078 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.335878] env[63531]: DEBUG oslo_vmware.api [None req-293635f3-1f0c-487a-ad12-e4a71b80c8dd tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118260, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.336994] env[63531]: DEBUG oslo_vmware.api [None req-04a987cd-eb1f-407a-8135-fd56b4798069 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Waiting for the task: (returnval){ [ 950.336994] env[63531]: value = "task-1118261" [ 950.336994] env[63531]: _type = "Task" [ 950.336994] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.344515] env[63531]: DEBUG oslo_vmware.api [None req-04a987cd-eb1f-407a-8135-fd56b4798069 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Task: {'id': task-1118261, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.507782] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a4bb48b8-dfee-4f65-99b6-ae50dabf2571 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.577559] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3ac1abf3-9cc3-459d-89d6-e0be3bbf01b0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "interface-a090135f-489d-457d-be36-ba6f61b71ab8-1d5fea6d-10c4-4740-81a1-8e1ac927dad5" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.191s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.645870] env[63531]: DEBUG oslo_vmware.api [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]526a9722-1822-fba7-c11a-1e853a1b1a94, 'name': SearchDatastore_Task, 'duration_secs': 0.010326} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.646179] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 950.646320] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 911f23f7-e320-48f3-87a9-0239b013feff/911f23f7-e320-48f3-87a9-0239b013feff.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 950.646579] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-37787f83-4f80-47ef-b481-50b3e170d368 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.652623] env[63531]: DEBUG oslo_vmware.api [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 950.652623] env[63531]: value = "task-1118262" [ 950.652623] env[63531]: _type = "Task" [ 950.652623] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.660439] env[63531]: DEBUG oslo_vmware.api [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118262, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.745858] env[63531]: DEBUG nova.scheduler.client.report [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 950.836125] env[63531]: DEBUG oslo_vmware.api [None req-293635f3-1f0c-487a-ad12-e4a71b80c8dd tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118260, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.16859} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.836423] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-293635f3-1f0c-487a-ad12-e4a71b80c8dd tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 950.836615] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-293635f3-1f0c-487a-ad12-e4a71b80c8dd tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 950.836796] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-293635f3-1f0c-487a-ad12-e4a71b80c8dd tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 950.836976] env[63531]: INFO nova.compute.manager [None req-293635f3-1f0c-487a-ad12-e4a71b80c8dd tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Took 1.19 seconds to destroy the instance on the hypervisor. [ 950.837249] env[63531]: DEBUG oslo.service.loopingcall [None req-293635f3-1f0c-487a-ad12-e4a71b80c8dd tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 950.837509] env[63531]: DEBUG nova.compute.manager [-] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 950.837606] env[63531]: DEBUG nova.network.neutron [-] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 950.848391] env[63531]: DEBUG oslo_vmware.api [None req-04a987cd-eb1f-407a-8135-fd56b4798069 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Task: {'id': task-1118261, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.146725} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.848624] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-04a987cd-eb1f-407a-8135-fd56b4798069 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 950.848806] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-04a987cd-eb1f-407a-8135-fd56b4798069 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] [instance: 55111b87-1e9a-4877-8689-987faa72a54e] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 950.848986] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-04a987cd-eb1f-407a-8135-fd56b4798069 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] [instance: 55111b87-1e9a-4877-8689-987faa72a54e] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 950.849177] env[63531]: INFO nova.compute.manager [None req-04a987cd-eb1f-407a-8135-fd56b4798069 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] [instance: 55111b87-1e9a-4877-8689-987faa72a54e] Took 1.20 seconds to destroy the instance on the hypervisor. [ 950.849410] env[63531]: DEBUG oslo.service.loopingcall [None req-04a987cd-eb1f-407a-8135-fd56b4798069 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 950.849600] env[63531]: DEBUG nova.compute.manager [-] [instance: 55111b87-1e9a-4877-8689-987faa72a54e] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 950.849696] env[63531]: DEBUG nova.network.neutron [-] [instance: 55111b87-1e9a-4877-8689-987faa72a54e] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 951.170471] env[63531]: DEBUG oslo_vmware.api [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118262, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.245267] env[63531]: DEBUG nova.compute.manager [req-b13df409-7fd5-4668-a0d3-4c0e5b6b2e59 req-9458ff07-ee27-4950-85ca-2a2c792d0803 service nova] [instance: 55111b87-1e9a-4877-8689-987faa72a54e] Received event network-vif-deleted-c2e8b778-965b-40b8-a5c7-8536e055435e {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 951.245508] env[63531]: INFO nova.compute.manager [req-b13df409-7fd5-4668-a0d3-4c0e5b6b2e59 req-9458ff07-ee27-4950-85ca-2a2c792d0803 service nova] [instance: 55111b87-1e9a-4877-8689-987faa72a54e] Neutron deleted interface c2e8b778-965b-40b8-a5c7-8536e055435e; detaching it from the instance and deleting it from the info cache [ 951.245725] env[63531]: DEBUG nova.network.neutron [req-b13df409-7fd5-4668-a0d3-4c0e5b6b2e59 req-9458ff07-ee27-4950-85ca-2a2c792d0803 service nova] [instance: 55111b87-1e9a-4877-8689-987faa72a54e] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.254835] env[63531]: DEBUG oslo_concurrency.lockutils [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.379s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.256119] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ff3c91f0-14af-4f9f-8d99-3fe699fd3d7c tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 18.362s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.292944] env[63531]: INFO nova.network.neutron [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Updating port a70530fb-7805-40a0-bea5-a2a666dc679d with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 951.618669] env[63531]: DEBUG nova.network.neutron [-] [instance: 55111b87-1e9a-4877-8689-987faa72a54e] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.665304] env[63531]: DEBUG oslo_vmware.api [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118262, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.53381} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.665641] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 911f23f7-e320-48f3-87a9-0239b013feff/911f23f7-e320-48f3-87a9-0239b013feff.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 951.665948] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 951.666259] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-40a48fcc-7267-4701-a4b6-ed02f0fffd6f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.672720] env[63531]: DEBUG oslo_vmware.api [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 951.672720] env[63531]: value = "task-1118263" [ 951.672720] env[63531]: _type = "Task" [ 951.672720] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.682020] env[63531]: DEBUG oslo_vmware.api [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118263, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.750030] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-00546a6c-ec9e-499e-952f-20203afea327 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.759868] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62f08f76-99df-40c3-a3fb-1bec0ad46099 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.800516] env[63531]: DEBUG nova.compute.manager [req-b13df409-7fd5-4668-a0d3-4c0e5b6b2e59 req-9458ff07-ee27-4950-85ca-2a2c792d0803 service nova] [instance: 55111b87-1e9a-4877-8689-987faa72a54e] Detach interface failed, port_id=c2e8b778-965b-40b8-a5c7-8536e055435e, reason: Instance 55111b87-1e9a-4877-8689-987faa72a54e could not be found. {{(pid=63531) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 952.026237] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c68a272-32be-4e6e-beea-a4b91b303134 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.033403] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64e686b6-fce2-4453-acd6-d85567fe6375 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.063934] env[63531]: DEBUG nova.network.neutron [-] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 952.065052] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19fc5920-87b3-4cf8-97ce-ddfd83da8f06 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.072484] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-451fcc03-54ab-4e54-a70c-33b1dda3cb31 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.086539] env[63531]: DEBUG nova.compute.provider_tree [None req-ff3c91f0-14af-4f9f-8d99-3fe699fd3d7c tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 952.122289] env[63531]: INFO nova.compute.manager [-] [instance: 55111b87-1e9a-4877-8689-987faa72a54e] Took 1.27 seconds to deallocate network for instance. [ 952.183315] env[63531]: DEBUG oslo_vmware.api [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118263, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066697} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.183602] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 952.184371] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21adea2d-2816-4c3d-9e4c-b189c79975c4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.206278] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Reconfiguring VM instance instance-00000055 to attach disk [datastore1] 911f23f7-e320-48f3-87a9-0239b013feff/911f23f7-e320-48f3-87a9-0239b013feff.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 952.207481] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3406f7c5-a700-466f-bc4f-d1cbf7f80fbd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.222065] env[63531]: DEBUG nova.compute.manager [req-2680ccad-a5ef-46db-8c60-bb0edbd69de8 req-845670cd-dac3-48d3-a727-93e02730e02d service nova] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Received event network-vif-deleted-17da826c-920c-4044-93cb-be8c8cf01c13 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 952.228563] env[63531]: DEBUG oslo_vmware.api [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 952.228563] env[63531]: value = "task-1118264" [ 952.228563] env[63531]: _type = "Task" [ 952.228563] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.237181] env[63531]: DEBUG oslo_vmware.api [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118264, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.569332] env[63531]: INFO nova.compute.manager [-] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Took 1.73 seconds to deallocate network for instance. [ 952.589814] env[63531]: DEBUG nova.scheduler.client.report [None req-ff3c91f0-14af-4f9f-8d99-3fe699fd3d7c tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 952.628995] env[63531]: DEBUG oslo_concurrency.lockutils [None req-04a987cd-eb1f-407a-8135-fd56b4798069 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.741380] env[63531]: DEBUG oslo_vmware.api [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118264, 'name': ReconfigVM_Task, 'duration_secs': 0.298751} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.741697] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Reconfigured VM instance instance-00000055 to attach disk [datastore1] 911f23f7-e320-48f3-87a9-0239b013feff/911f23f7-e320-48f3-87a9-0239b013feff.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 952.742268] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8e3d177b-ebaa-4587-9905-c2c3c1c69a4e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.749445] env[63531]: DEBUG oslo_vmware.api [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 952.749445] env[63531]: value = "task-1118265" [ 952.749445] env[63531]: _type = "Task" [ 952.749445] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.757371] env[63531]: DEBUG oslo_vmware.api [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118265, 'name': Rename_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.772362] env[63531]: DEBUG oslo_concurrency.lockutils [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Acquiring lock "refresh_cache-1ca7b851-2a06-4181-8271-58aafcd322d6" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.772641] env[63531]: DEBUG oslo_concurrency.lockutils [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Acquired lock "refresh_cache-1ca7b851-2a06-4181-8271-58aafcd322d6" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.772763] env[63531]: DEBUG nova.network.neutron [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 953.079855] env[63531]: DEBUG oslo_concurrency.lockutils [None req-293635f3-1f0c-487a-ad12-e4a71b80c8dd tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.259643] env[63531]: DEBUG oslo_vmware.api [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118265, 'name': Rename_Task, 'duration_secs': 0.136489} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.259942] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 953.260766] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-00fcc5ef-4e9f-4d29-bce3-d0e62477c4af {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.266272] env[63531]: DEBUG oslo_vmware.api [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 953.266272] env[63531]: value = "task-1118266" [ 953.266272] env[63531]: _type = "Task" [ 953.266272] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.273864] env[63531]: DEBUG oslo_vmware.api [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118266, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.278206] env[63531]: DEBUG nova.compute.manager [req-188cdf7e-9772-45a7-accb-c10ba76ef011 req-9449cbed-7311-4ee7-916c-1c4d3de775a5 service nova] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Received event network-vif-plugged-a70530fb-7805-40a0-bea5-a2a666dc679d {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 953.278298] env[63531]: DEBUG oslo_concurrency.lockutils [req-188cdf7e-9772-45a7-accb-c10ba76ef011 req-9449cbed-7311-4ee7-916c-1c4d3de775a5 service nova] Acquiring lock "1ca7b851-2a06-4181-8271-58aafcd322d6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.278479] env[63531]: DEBUG oslo_concurrency.lockutils [req-188cdf7e-9772-45a7-accb-c10ba76ef011 req-9449cbed-7311-4ee7-916c-1c4d3de775a5 service nova] Lock "1ca7b851-2a06-4181-8271-58aafcd322d6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.278652] env[63531]: DEBUG oslo_concurrency.lockutils [req-188cdf7e-9772-45a7-accb-c10ba76ef011 req-9449cbed-7311-4ee7-916c-1c4d3de775a5 service nova] Lock "1ca7b851-2a06-4181-8271-58aafcd322d6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.278852] env[63531]: DEBUG nova.compute.manager [req-188cdf7e-9772-45a7-accb-c10ba76ef011 req-9449cbed-7311-4ee7-916c-1c4d3de775a5 service nova] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] No waiting events found dispatching network-vif-plugged-a70530fb-7805-40a0-bea5-a2a666dc679d {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 953.279062] env[63531]: WARNING nova.compute.manager [req-188cdf7e-9772-45a7-accb-c10ba76ef011 req-9449cbed-7311-4ee7-916c-1c4d3de775a5 service nova] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Received unexpected event network-vif-plugged-a70530fb-7805-40a0-bea5-a2a666dc679d for instance with vm_state shelved_offloaded and task_state spawning. [ 953.279228] env[63531]: DEBUG nova.compute.manager [req-188cdf7e-9772-45a7-accb-c10ba76ef011 req-9449cbed-7311-4ee7-916c-1c4d3de775a5 service nova] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Received event network-changed-a70530fb-7805-40a0-bea5-a2a666dc679d {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 953.279356] env[63531]: DEBUG nova.compute.manager [req-188cdf7e-9772-45a7-accb-c10ba76ef011 req-9449cbed-7311-4ee7-916c-1c4d3de775a5 service nova] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Refreshing instance network info cache due to event network-changed-a70530fb-7805-40a0-bea5-a2a666dc679d. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 953.279545] env[63531]: DEBUG oslo_concurrency.lockutils [req-188cdf7e-9772-45a7-accb-c10ba76ef011 req-9449cbed-7311-4ee7-916c-1c4d3de775a5 service nova] Acquiring lock "refresh_cache-1ca7b851-2a06-4181-8271-58aafcd322d6" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.472657] env[63531]: DEBUG nova.network.neutron [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Updating instance_info_cache with network_info: [{"id": "a70530fb-7805-40a0-bea5-a2a666dc679d", "address": "fa:16:3e:78:c3:90", "network": {"id": "73e6d3ed-2016-498a-bda6-fd7dfbff93da", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1501868918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f727ddb992dc4c06954f25f6a1613b4a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c20f5114-0866-45b3-9a7c-62f113ff83fa", "external-id": "nsx-vlan-transportzone-47", "segmentation_id": 47, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa70530fb-78", "ovs_interfaceid": "a70530fb-7805-40a0-bea5-a2a666dc679d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 953.601683] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ff3c91f0-14af-4f9f-8d99-3fe699fd3d7c tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.346s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.601928] env[63531]: DEBUG nova.compute.manager [None req-ff3c91f0-14af-4f9f-8d99-3fe699fd3d7c tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Resized/migrated instance is powered off. Setting vm_state to 'stopped'. {{(pid=63531) _confirm_resize /opt/stack/nova/nova/compute/manager.py:4909}} [ 953.605323] env[63531]: DEBUG oslo_concurrency.lockutils [None req-141fc11b-5904-4c16-af8a-e3b6485bc843 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.490s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.605544] env[63531]: DEBUG oslo_concurrency.lockutils [None req-141fc11b-5904-4c16-af8a-e3b6485bc843 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.607932] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b9944abd-dda8-40ee-b4da-5c9bb59d9201 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.772s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.608162] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b9944abd-dda8-40ee-b4da-5c9bb59d9201 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.609974] env[63531]: DEBUG oslo_concurrency.lockutils [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 14.662s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.610142] env[63531]: DEBUG nova.objects.instance [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63531) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 953.634129] env[63531]: INFO nova.scheduler.client.report [None req-b9944abd-dda8-40ee-b4da-5c9bb59d9201 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Deleted allocations for instance 6118e242-c0db-4707-a5f9-3d12b823935f [ 953.635944] env[63531]: INFO nova.scheduler.client.report [None req-141fc11b-5904-4c16-af8a-e3b6485bc843 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Deleted allocations for instance 4b5e5821-9def-4b53-be19-355d9e5f81ec [ 953.775622] env[63531]: DEBUG oslo_vmware.api [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118266, 'name': PowerOnVM_Task, 'duration_secs': 0.431195} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.775907] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 953.776134] env[63531]: INFO nova.compute.manager [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Took 8.23 seconds to spawn the instance on the hypervisor. [ 953.776321] env[63531]: DEBUG nova.compute.manager [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 953.777065] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51375c57-2bac-4c04-9018-5169d5039521 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.975680] env[63531]: DEBUG oslo_concurrency.lockutils [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Releasing lock "refresh_cache-1ca7b851-2a06-4181-8271-58aafcd322d6" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.979183] env[63531]: DEBUG oslo_concurrency.lockutils [req-188cdf7e-9772-45a7-accb-c10ba76ef011 req-9449cbed-7311-4ee7-916c-1c4d3de775a5 service nova] Acquired lock "refresh_cache-1ca7b851-2a06-4181-8271-58aafcd322d6" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.979431] env[63531]: DEBUG nova.network.neutron [req-188cdf7e-9772-45a7-accb-c10ba76ef011 req-9449cbed-7311-4ee7-916c-1c4d3de775a5 service nova] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Refreshing network info cache for port a70530fb-7805-40a0-bea5-a2a666dc679d {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 954.004929] env[63531]: DEBUG nova.virt.hardware [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='e4371820df43c9fd1ff1720add495e4f',container_format='bare',created_at=2024-10-03T08:21:56Z,direct_url=,disk_format='vmdk',id=d2fca167-8767-4b16-b64a-61fd11e58a0e,min_disk=1,min_ram=0,name='tempest-ServersNegativeTestJSON-server-578555969-shelved',owner='f727ddb992dc4c06954f25f6a1613b4a',properties=ImageMetaProps,protected=,size=31666176,status='active',tags=,updated_at=2024-10-03T08:22:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 954.005236] env[63531]: DEBUG nova.virt.hardware [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 954.005445] env[63531]: DEBUG nova.virt.hardware [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 954.005728] env[63531]: DEBUG nova.virt.hardware [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 954.005854] env[63531]: DEBUG nova.virt.hardware [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 954.006112] env[63531]: DEBUG nova.virt.hardware [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 954.006292] env[63531]: DEBUG nova.virt.hardware [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 954.006482] env[63531]: DEBUG nova.virt.hardware [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 954.006697] env[63531]: DEBUG nova.virt.hardware [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 954.006882] env[63531]: DEBUG nova.virt.hardware [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 954.007077] env[63531]: DEBUG nova.virt.hardware [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 954.008660] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1308d60c-5260-4e28-b3cb-7b0788df235e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.018671] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e40896d-4dd1-40ee-adb4-d11070f4f1b5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.031971] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:78:c3:90', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c20f5114-0866-45b3-9a7c-62f113ff83fa', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a70530fb-7805-40a0-bea5-a2a666dc679d', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 954.039825] env[63531]: DEBUG oslo.service.loopingcall [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 954.040121] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 954.040312] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ce591b20-6543-401a-a992-e8b0f5f2248c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.060535] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 954.060535] env[63531]: value = "task-1118267" [ 954.060535] env[63531]: _type = "Task" [ 954.060535] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.067705] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118267, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.145017] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b9944abd-dda8-40ee-b4da-5c9bb59d9201 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "6118e242-c0db-4707-a5f9-3d12b823935f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.836s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.149021] env[63531]: DEBUG oslo_concurrency.lockutils [None req-141fc11b-5904-4c16-af8a-e3b6485bc843 tempest-ImagesTestJSON-1303287413 tempest-ImagesTestJSON-1303287413-project-member] Lock "4b5e5821-9def-4b53-be19-355d9e5f81ec" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.046s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.176042] env[63531]: INFO nova.scheduler.client.report [None req-ff3c91f0-14af-4f9f-8d99-3fe699fd3d7c tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Deleted allocation for migration 72374b80-67c2-4204-8fdb-7c6eff83d384 [ 954.295718] env[63531]: INFO nova.compute.manager [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Took 27.28 seconds to build instance. [ 954.489010] env[63531]: DEBUG nova.objects.instance [None req-cadfeb91-2514-4ff0-82ee-3a9286d46ee1 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lazy-loading 'flavor' on Instance uuid 9a3fad46-a15a-451c-bdab-a3c8cc8add07 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 954.572011] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118267, 'name': CreateVM_Task, 'duration_secs': 0.315974} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.572205] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 954.572896] env[63531]: DEBUG oslo_concurrency.lockutils [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d2fca167-8767-4b16-b64a-61fd11e58a0e" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 954.573084] env[63531]: DEBUG oslo_concurrency.lockutils [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d2fca167-8767-4b16-b64a-61fd11e58a0e" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.573478] env[63531]: DEBUG oslo_concurrency.lockutils [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d2fca167-8767-4b16-b64a-61fd11e58a0e" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 954.573746] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b5456094-e82d-43c5-967c-24969b83bc7b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.580634] env[63531]: DEBUG oslo_vmware.api [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Waiting for the task: (returnval){ [ 954.580634] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52988b66-f70d-7596-e855-e0ff66d47552" [ 954.580634] env[63531]: _type = "Task" [ 954.580634] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.587423] env[63531]: DEBUG oslo_vmware.api [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52988b66-f70d-7596-e855-e0ff66d47552, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.624645] env[63531]: DEBUG oslo_concurrency.lockutils [None req-90f0049e-84c7-431a-b144-44ebd3711b61 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.015s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.625814] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.319s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.627315] env[63531]: INFO nova.compute.claims [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 954.684884] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ff3c91f0-14af-4f9f-8d99-3fe699fd3d7c tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "9a3fad46-a15a-451c-bdab-a3c8cc8add07" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 25.221s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.797291] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e012e76e-a3b9-4263-8e7d-1a041d2a6abb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "911f23f7-e320-48f3-87a9-0239b013feff" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.795s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.847728] env[63531]: DEBUG nova.network.neutron [req-188cdf7e-9772-45a7-accb-c10ba76ef011 req-9449cbed-7311-4ee7-916c-1c4d3de775a5 service nova] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Updated VIF entry in instance network info cache for port a70530fb-7805-40a0-bea5-a2a666dc679d. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 954.848492] env[63531]: DEBUG nova.network.neutron [req-188cdf7e-9772-45a7-accb-c10ba76ef011 req-9449cbed-7311-4ee7-916c-1c4d3de775a5 service nova] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Updating instance_info_cache with network_info: [{"id": "a70530fb-7805-40a0-bea5-a2a666dc679d", "address": "fa:16:3e:78:c3:90", "network": {"id": "73e6d3ed-2016-498a-bda6-fd7dfbff93da", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1501868918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f727ddb992dc4c06954f25f6a1613b4a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c20f5114-0866-45b3-9a7c-62f113ff83fa", "external-id": "nsx-vlan-transportzone-47", "segmentation_id": 47, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa70530fb-78", "ovs_interfaceid": "a70530fb-7805-40a0-bea5-a2a666dc679d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.997413] env[63531]: DEBUG oslo_concurrency.lockutils [None req-cadfeb91-2514-4ff0-82ee-3a9286d46ee1 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquiring lock "refresh_cache-9a3fad46-a15a-451c-bdab-a3c8cc8add07" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 954.997565] env[63531]: DEBUG oslo_concurrency.lockutils [None req-cadfeb91-2514-4ff0-82ee-3a9286d46ee1 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquired lock "refresh_cache-9a3fad46-a15a-451c-bdab-a3c8cc8add07" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.997638] env[63531]: DEBUG nova.network.neutron [None req-cadfeb91-2514-4ff0-82ee-3a9286d46ee1 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 954.997890] env[63531]: DEBUG nova.objects.instance [None req-cadfeb91-2514-4ff0-82ee-3a9286d46ee1 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lazy-loading 'info_cache' on Instance uuid 9a3fad46-a15a-451c-bdab-a3c8cc8add07 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 955.091184] env[63531]: DEBUG oslo_concurrency.lockutils [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d2fca167-8767-4b16-b64a-61fd11e58a0e" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.091409] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Processing image d2fca167-8767-4b16-b64a-61fd11e58a0e {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 955.091660] env[63531]: DEBUG oslo_concurrency.lockutils [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d2fca167-8767-4b16-b64a-61fd11e58a0e/d2fca167-8767-4b16-b64a-61fd11e58a0e.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.091816] env[63531]: DEBUG oslo_concurrency.lockutils [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d2fca167-8767-4b16-b64a-61fd11e58a0e/d2fca167-8767-4b16-b64a-61fd11e58a0e.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.092012] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 955.092275] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3a5243f0-3688-49e8-911c-b58ed17e3930 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.100527] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 955.100711] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 955.101436] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-013cc529-30c5-410f-af35-01cfb753574d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.106871] env[63531]: DEBUG oslo_vmware.api [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Waiting for the task: (returnval){ [ 955.106871] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]520b1ed3-433f-fed5-08c0-4674fa428215" [ 955.106871] env[63531]: _type = "Task" [ 955.106871] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.114627] env[63531]: DEBUG oslo_vmware.api [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]520b1ed3-433f-fed5-08c0-4674fa428215, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.351816] env[63531]: DEBUG oslo_concurrency.lockutils [req-188cdf7e-9772-45a7-accb-c10ba76ef011 req-9449cbed-7311-4ee7-916c-1c4d3de775a5 service nova] Releasing lock "refresh_cache-1ca7b851-2a06-4181-8271-58aafcd322d6" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.502184] env[63531]: DEBUG nova.objects.base [None req-cadfeb91-2514-4ff0-82ee-3a9286d46ee1 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Object Instance<9a3fad46-a15a-451c-bdab-a3c8cc8add07> lazy-loaded attributes: flavor,info_cache {{(pid=63531) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 955.565258] env[63531]: DEBUG nova.compute.manager [req-b6d2a985-a8c5-4b24-a4a9-040390c341e7 req-f6da27fd-3c38-414e-99d6-79637b35e1a5 service nova] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Received event network-changed-d3f47a79-f699-4d63-9412-724fe1f15913 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 955.565522] env[63531]: DEBUG nova.compute.manager [req-b6d2a985-a8c5-4b24-a4a9-040390c341e7 req-f6da27fd-3c38-414e-99d6-79637b35e1a5 service nova] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Refreshing instance network info cache due to event network-changed-d3f47a79-f699-4d63-9412-724fe1f15913. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 955.565889] env[63531]: DEBUG oslo_concurrency.lockutils [req-b6d2a985-a8c5-4b24-a4a9-040390c341e7 req-f6da27fd-3c38-414e-99d6-79637b35e1a5 service nova] Acquiring lock "refresh_cache-911f23f7-e320-48f3-87a9-0239b013feff" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.566108] env[63531]: DEBUG oslo_concurrency.lockutils [req-b6d2a985-a8c5-4b24-a4a9-040390c341e7 req-f6da27fd-3c38-414e-99d6-79637b35e1a5 service nova] Acquired lock "refresh_cache-911f23f7-e320-48f3-87a9-0239b013feff" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.566336] env[63531]: DEBUG nova.network.neutron [req-b6d2a985-a8c5-4b24-a4a9-040390c341e7 req-f6da27fd-3c38-414e-99d6-79637b35e1a5 service nova] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Refreshing network info cache for port d3f47a79-f699-4d63-9412-724fe1f15913 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 955.619992] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Preparing fetch location {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 955.620314] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Fetch image to [datastore1] OSTACK_IMG_3e708714-2e59-4e1b-bd66-ad64677b1d1d/OSTACK_IMG_3e708714-2e59-4e1b-bd66-ad64677b1d1d.vmdk {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 955.620524] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Downloading stream optimized image d2fca167-8767-4b16-b64a-61fd11e58a0e to [datastore1] OSTACK_IMG_3e708714-2e59-4e1b-bd66-ad64677b1d1d/OSTACK_IMG_3e708714-2e59-4e1b-bd66-ad64677b1d1d.vmdk on the data store datastore1 as vApp {{(pid=63531) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 955.620715] env[63531]: DEBUG nova.virt.vmwareapi.images [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Downloading image file data d2fca167-8767-4b16-b64a-61fd11e58a0e to the ESX as VM named 'OSTACK_IMG_3e708714-2e59-4e1b-bd66-ad64677b1d1d' {{(pid=63531) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 955.703210] env[63531]: DEBUG oslo_vmware.rw_handles [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 955.703210] env[63531]: value = "resgroup-9" [ 955.703210] env[63531]: _type = "ResourcePool" [ 955.703210] env[63531]: }. {{(pid=63531) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 955.703874] env[63531]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-5b1a35d0-b075-49a1-a667-a64779d1c2b2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.728501] env[63531]: DEBUG oslo_vmware.rw_handles [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Lease: (returnval){ [ 955.728501] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5256ecb5-cb63-1fd6-7e85-924bb3215c14" [ 955.728501] env[63531]: _type = "HttpNfcLease" [ 955.728501] env[63531]: } obtained for vApp import into resource pool (val){ [ 955.728501] env[63531]: value = "resgroup-9" [ 955.728501] env[63531]: _type = "ResourcePool" [ 955.728501] env[63531]: }. {{(pid=63531) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 955.729269] env[63531]: DEBUG oslo_vmware.api [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Waiting for the lease: (returnval){ [ 955.729269] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5256ecb5-cb63-1fd6-7e85-924bb3215c14" [ 955.729269] env[63531]: _type = "HttpNfcLease" [ 955.729269] env[63531]: } to be ready. {{(pid=63531) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 955.739670] env[63531]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 955.739670] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5256ecb5-cb63-1fd6-7e85-924bb3215c14" [ 955.739670] env[63531]: _type = "HttpNfcLease" [ 955.739670] env[63531]: } is initializing. {{(pid=63531) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 955.807893] env[63531]: DEBUG nova.compute.manager [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Stashing vm_state: active {{(pid=63531) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 955.979632] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7e2d9f1-79d9-4682-b6db-a7bf4fb10523 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.990366] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-964871f5-2fc8-4a19-9542-2f43cd1c32e6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.023104] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-156cb801-fbdf-49c8-b939-af13813aecb9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.031255] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-447ea116-c275-4070-b44d-c21ec483e6a1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.045173] env[63531]: DEBUG nova.compute.provider_tree [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 956.238490] env[63531]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 956.238490] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5256ecb5-cb63-1fd6-7e85-924bb3215c14" [ 956.238490] env[63531]: _type = "HttpNfcLease" [ 956.238490] env[63531]: } is initializing. {{(pid=63531) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 956.338939] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.393490] env[63531]: DEBUG nova.network.neutron [None req-cadfeb91-2514-4ff0-82ee-3a9286d46ee1 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Updating instance_info_cache with network_info: [{"id": "e4133e0f-55c1-4b3f-847d-899dd1167eec", "address": "fa:16:3e:90:40:9b", "network": {"id": "c4f4c994-36ab-42ab-94f8-78f29cbafa50", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-994507698-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "57e9bf4ee0f1463fa67b0f8deff30fdc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "767a3a48-41d4-4a0c-961d-0024837f63bd", "external-id": "nsx-vlan-transportzone-141", "segmentation_id": 141, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4133e0f-55", "ovs_interfaceid": "e4133e0f-55c1-4b3f-847d-899dd1167eec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.412712] env[63531]: DEBUG nova.network.neutron [req-b6d2a985-a8c5-4b24-a4a9-040390c341e7 req-f6da27fd-3c38-414e-99d6-79637b35e1a5 service nova] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Updated VIF entry in instance network info cache for port d3f47a79-f699-4d63-9412-724fe1f15913. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 956.413098] env[63531]: DEBUG nova.network.neutron [req-b6d2a985-a8c5-4b24-a4a9-040390c341e7 req-f6da27fd-3c38-414e-99d6-79637b35e1a5 service nova] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Updating instance_info_cache with network_info: [{"id": "d3f47a79-f699-4d63-9412-724fe1f15913", "address": "fa:16:3e:71:30:b9", "network": {"id": "eac8786d-bc67-4916-8483-b53d9a25efb8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1604989690-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.206", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "abdae960d43b47bc95ab41f265d2c85b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db068f71-08cc-42d4-8ab6-17134c1585e5", "external-id": "nsx-vlan-transportzone-721", "segmentation_id": 721, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd3f47a79-f6", "ovs_interfaceid": "d3f47a79-f699-4d63-9412-724fe1f15913", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.548515] env[63531]: DEBUG nova.scheduler.client.report [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 956.742690] env[63531]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 956.742690] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5256ecb5-cb63-1fd6-7e85-924bb3215c14" [ 956.742690] env[63531]: _type = "HttpNfcLease" [ 956.742690] env[63531]: } is ready. {{(pid=63531) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 956.743058] env[63531]: DEBUG oslo_vmware.rw_handles [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 956.743058] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5256ecb5-cb63-1fd6-7e85-924bb3215c14" [ 956.743058] env[63531]: _type = "HttpNfcLease" [ 956.743058] env[63531]: }. {{(pid=63531) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 956.745693] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34141720-04a1-45f5-88a8-00980df90d95 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.752547] env[63531]: DEBUG oslo_vmware.rw_handles [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52835f70-f509-a892-31e5-9932e4f2593f/disk-0.vmdk from lease info. {{(pid=63531) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 956.752756] env[63531]: DEBUG oslo_vmware.rw_handles [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Creating HTTP connection to write to file with size = 31666176 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52835f70-f509-a892-31e5-9932e4f2593f/disk-0.vmdk. {{(pid=63531) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 956.823062] env[63531]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-d8bb8e31-77d1-4589-990d-7256f377ed56 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.897312] env[63531]: DEBUG oslo_concurrency.lockutils [None req-cadfeb91-2514-4ff0-82ee-3a9286d46ee1 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Releasing lock "refresh_cache-9a3fad46-a15a-451c-bdab-a3c8cc8add07" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 956.916183] env[63531]: DEBUG oslo_concurrency.lockutils [req-b6d2a985-a8c5-4b24-a4a9-040390c341e7 req-f6da27fd-3c38-414e-99d6-79637b35e1a5 service nova] Releasing lock "refresh_cache-911f23f7-e320-48f3-87a9-0239b013feff" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 957.054259] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.428s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.054813] env[63531]: DEBUG nova.compute.manager [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 957.058078] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4c36b677-1dba-4034-8c3f-d7d2eda1da73 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.010s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.058284] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4c36b677-1dba-4034-8c3f-d7d2eda1da73 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.064459] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ba2842a1-0c0c-4cb4-b87b-242607074b48 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.303s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.064695] env[63531]: DEBUG nova.objects.instance [None req-ba2842a1-0c0c-4cb4-b87b-242607074b48 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lazy-loading 'resources' on Instance uuid 3fb0499d-8074-4e63-be9f-380730416cc4 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 957.107551] env[63531]: INFO nova.scheduler.client.report [None req-4c36b677-1dba-4034-8c3f-d7d2eda1da73 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Deleted allocations for instance e065f052-ba3e-4783-8953-5dc200d1f3e9 [ 957.399483] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-cadfeb91-2514-4ff0-82ee-3a9286d46ee1 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 957.399890] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9f55531a-053b-408c-ae67-ff179cb04eb3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.407611] env[63531]: DEBUG oslo_vmware.api [None req-cadfeb91-2514-4ff0-82ee-3a9286d46ee1 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 957.407611] env[63531]: value = "task-1118269" [ 957.407611] env[63531]: _type = "Task" [ 957.407611] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.416407] env[63531]: DEBUG oslo_vmware.api [None req-cadfeb91-2514-4ff0-82ee-3a9286d46ee1 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118269, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.579560] env[63531]: DEBUG nova.compute.utils [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 957.586052] env[63531]: DEBUG nova.compute.manager [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 957.586324] env[63531]: DEBUG nova.network.neutron [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 957.592842] env[63531]: DEBUG nova.compute.manager [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 957.624688] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4c36b677-1dba-4034-8c3f-d7d2eda1da73 tempest-ServerShowV254Test-1322516443 tempest-ServerShowV254Test-1322516443-project-member] Lock "e065f052-ba3e-4783-8953-5dc200d1f3e9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.392s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.680801] env[63531]: DEBUG nova.policy [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0372c22d1e9e4f66a0417643dab4a2ec', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '94a95667c00947ea9183a6307c569c90', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 957.924572] env[63531]: DEBUG oslo_vmware.api [None req-cadfeb91-2514-4ff0-82ee-3a9286d46ee1 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118269, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.060092] env[63531]: DEBUG nova.network.neutron [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Successfully created port: 1bb5f5b0-d6c9-4b8d-be14-e6e98f2aeef1 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 958.069507] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36d0c5ca-45fa-476b-830a-213cfd4038d1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.079835] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-762fdfe3-116a-41c0-a092-cc60ca9e86f7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.137303] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b976c8c2-ed77-4527-b24c-7b0c9ac31cdd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.143022] env[63531]: DEBUG oslo_vmware.rw_handles [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Completed reading data from the image iterator. {{(pid=63531) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 958.143022] env[63531]: DEBUG oslo_vmware.rw_handles [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52835f70-f509-a892-31e5-9932e4f2593f/disk-0.vmdk. {{(pid=63531) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 958.143485] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95f4e2b2-a69d-448b-bff9-e748c4e2810b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.159721] env[63531]: DEBUG oslo_vmware.rw_handles [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52835f70-f509-a892-31e5-9932e4f2593f/disk-0.vmdk is in state: ready. {{(pid=63531) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 958.159721] env[63531]: DEBUG oslo_vmware.rw_handles [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52835f70-f509-a892-31e5-9932e4f2593f/disk-0.vmdk. {{(pid=63531) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 958.159721] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42079c17-c448-4481-885f-f3abfc55247c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.161254] env[63531]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-952ecb41-7ad0-488e-9156-cd453ff13422 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.176355] env[63531]: DEBUG nova.compute.provider_tree [None req-ba2842a1-0c0c-4cb4-b87b-242607074b48 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 958.424111] env[63531]: DEBUG oslo_vmware.api [None req-cadfeb91-2514-4ff0-82ee-3a9286d46ee1 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118269, 'name': PowerOnVM_Task, 'duration_secs': 0.582105} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.424111] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-cadfeb91-2514-4ff0-82ee-3a9286d46ee1 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 958.424587] env[63531]: DEBUG nova.compute.manager [None req-cadfeb91-2514-4ff0-82ee-3a9286d46ee1 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 958.426064] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfd14aae-c5c5-434a-b41e-84106257073b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.519355] env[63531]: DEBUG oslo_vmware.rw_handles [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52835f70-f509-a892-31e5-9932e4f2593f/disk-0.vmdk. {{(pid=63531) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 958.519355] env[63531]: INFO nova.virt.vmwareapi.images [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Downloaded image file data d2fca167-8767-4b16-b64a-61fd11e58a0e [ 958.519355] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e037008-2793-4524-b826-70287697fadc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.537414] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b1c89d93-43e0-4d0c-bc49-386edbfd0406 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.641510] env[63531]: DEBUG nova.compute.manager [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 958.666206] env[63531]: INFO nova.virt.vmwareapi.images [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] The imported VM was unregistered [ 958.669335] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Caching image {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 958.669616] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Creating directory with path [datastore1] devstack-image-cache_base/d2fca167-8767-4b16-b64a-61fd11e58a0e {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 958.671856] env[63531]: DEBUG nova.virt.hardware [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 958.672155] env[63531]: DEBUG nova.virt.hardware [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 958.672369] env[63531]: DEBUG nova.virt.hardware [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 958.672604] env[63531]: DEBUG nova.virt.hardware [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 958.672792] env[63531]: DEBUG nova.virt.hardware [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 958.673027] env[63531]: DEBUG nova.virt.hardware [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 958.673295] env[63531]: DEBUG nova.virt.hardware [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 958.673506] env[63531]: DEBUG nova.virt.hardware [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 958.673721] env[63531]: DEBUG nova.virt.hardware [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 958.673928] env[63531]: DEBUG nova.virt.hardware [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 958.674163] env[63531]: DEBUG nova.virt.hardware [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 958.674459] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-262c8b5e-7d21-4bac-bb0d-34e63d144d57 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.676993] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-027e27e3-9226-433c-8982-2edf38fd3493 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.680652] env[63531]: DEBUG nova.scheduler.client.report [None req-ba2842a1-0c0c-4cb4-b87b-242607074b48 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 958.689660] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab08f8f4-342d-4c57-84fd-ad526d1e6756 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.704676] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Created directory with path [datastore1] devstack-image-cache_base/d2fca167-8767-4b16-b64a-61fd11e58a0e {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 958.704838] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_3e708714-2e59-4e1b-bd66-ad64677b1d1d/OSTACK_IMG_3e708714-2e59-4e1b-bd66-ad64677b1d1d.vmdk to [datastore1] devstack-image-cache_base/d2fca167-8767-4b16-b64a-61fd11e58a0e/d2fca167-8767-4b16-b64a-61fd11e58a0e.vmdk. {{(pid=63531) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 958.705097] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-c964462e-85e2-44c8-9da2-248945082197 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.711736] env[63531]: DEBUG oslo_vmware.api [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Waiting for the task: (returnval){ [ 958.711736] env[63531]: value = "task-1118271" [ 958.711736] env[63531]: _type = "Task" [ 958.711736] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.719795] env[63531]: DEBUG oslo_vmware.api [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1118271, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.186983] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ba2842a1-0c0c-4cb4-b87b-242607074b48 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.122s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.191305] env[63531]: DEBUG oslo_concurrency.lockutils [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 12.013s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.223917] env[63531]: INFO nova.scheduler.client.report [None req-ba2842a1-0c0c-4cb4-b87b-242607074b48 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Deleted allocations for instance 3fb0499d-8074-4e63-be9f-380730416cc4 [ 959.231983] env[63531]: DEBUG oslo_vmware.api [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1118271, 'name': MoveVirtualDisk_Task} progress is 18%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.449528] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a384b8d7-53a1-4e7f-90b4-de586fca38e3 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquiring lock "9a3fad46-a15a-451c-bdab-a3c8cc8add07" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.449770] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a384b8d7-53a1-4e7f-90b4-de586fca38e3 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "9a3fad46-a15a-451c-bdab-a3c8cc8add07" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.449989] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a384b8d7-53a1-4e7f-90b4-de586fca38e3 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquiring lock "9a3fad46-a15a-451c-bdab-a3c8cc8add07-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.450233] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a384b8d7-53a1-4e7f-90b4-de586fca38e3 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "9a3fad46-a15a-451c-bdab-a3c8cc8add07-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.450365] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a384b8d7-53a1-4e7f-90b4-de586fca38e3 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "9a3fad46-a15a-451c-bdab-a3c8cc8add07-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.452853] env[63531]: INFO nova.compute.manager [None req-a384b8d7-53a1-4e7f-90b4-de586fca38e3 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Terminating instance [ 959.454849] env[63531]: DEBUG nova.compute.manager [None req-a384b8d7-53a1-4e7f-90b4-de586fca38e3 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 959.455096] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a384b8d7-53a1-4e7f-90b4-de586fca38e3 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 959.456426] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6a4506f-4945-452d-b07b-d734715338c8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.466335] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-a384b8d7-53a1-4e7f-90b4-de586fca38e3 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 959.466986] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b186482b-5474-446b-85ac-2b37a6765b96 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.478083] env[63531]: DEBUG oslo_vmware.api [None req-a384b8d7-53a1-4e7f-90b4-de586fca38e3 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 959.478083] env[63531]: value = "task-1118272" [ 959.478083] env[63531]: _type = "Task" [ 959.478083] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.490544] env[63531]: DEBUG oslo_vmware.api [None req-a384b8d7-53a1-4e7f-90b4-de586fca38e3 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118272, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.702144] env[63531]: INFO nova.compute.claims [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 959.728791] env[63531]: DEBUG oslo_vmware.api [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1118271, 'name': MoveVirtualDisk_Task} progress is 38%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.739528] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ba2842a1-0c0c-4cb4-b87b-242607074b48 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "3fb0499d-8074-4e63-be9f-380730416cc4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.925s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.991589] env[63531]: DEBUG oslo_vmware.api [None req-a384b8d7-53a1-4e7f-90b4-de586fca38e3 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118272, 'name': PowerOffVM_Task, 'duration_secs': 0.354542} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.991976] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-a384b8d7-53a1-4e7f-90b4-de586fca38e3 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 959.992883] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a384b8d7-53a1-4e7f-90b4-de586fca38e3 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 959.993258] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ae81c0f4-8799-430f-a8e9-0069abbddd87 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.072056] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a384b8d7-53a1-4e7f-90b4-de586fca38e3 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 960.072056] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a384b8d7-53a1-4e7f-90b4-de586fca38e3 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 960.072056] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-a384b8d7-53a1-4e7f-90b4-de586fca38e3 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Deleting the datastore file [datastore1] 9a3fad46-a15a-451c-bdab-a3c8cc8add07 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 960.072056] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-28976420-848d-496b-af00-5514496a86ad {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.080516] env[63531]: DEBUG oslo_vmware.api [None req-a384b8d7-53a1-4e7f-90b4-de586fca38e3 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 960.080516] env[63531]: value = "task-1118274" [ 960.080516] env[63531]: _type = "Task" [ 960.080516] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.091829] env[63531]: DEBUG oslo_vmware.api [None req-a384b8d7-53a1-4e7f-90b4-de586fca38e3 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118274, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.157704] env[63531]: DEBUG nova.network.neutron [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Successfully updated port: 1bb5f5b0-d6c9-4b8d-be14-e6e98f2aeef1 {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 960.164962] env[63531]: DEBUG nova.compute.manager [req-e988c3e5-ea5f-4b3e-a0d1-c03c460b5c77 req-aeeee0fd-3af1-4cbd-884d-4eb2d433101a service nova] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Received event network-vif-plugged-1bb5f5b0-d6c9-4b8d-be14-e6e98f2aeef1 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 960.165740] env[63531]: DEBUG oslo_concurrency.lockutils [req-e988c3e5-ea5f-4b3e-a0d1-c03c460b5c77 req-aeeee0fd-3af1-4cbd-884d-4eb2d433101a service nova] Acquiring lock "6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.166098] env[63531]: DEBUG oslo_concurrency.lockutils [req-e988c3e5-ea5f-4b3e-a0d1-c03c460b5c77 req-aeeee0fd-3af1-4cbd-884d-4eb2d433101a service nova] Lock "6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.167244] env[63531]: DEBUG oslo_concurrency.lockutils [req-e988c3e5-ea5f-4b3e-a0d1-c03c460b5c77 req-aeeee0fd-3af1-4cbd-884d-4eb2d433101a service nova] Lock "6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.167244] env[63531]: DEBUG nova.compute.manager [req-e988c3e5-ea5f-4b3e-a0d1-c03c460b5c77 req-aeeee0fd-3af1-4cbd-884d-4eb2d433101a service nova] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] No waiting events found dispatching network-vif-plugged-1bb5f5b0-d6c9-4b8d-be14-e6e98f2aeef1 {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 960.167244] env[63531]: WARNING nova.compute.manager [req-e988c3e5-ea5f-4b3e-a0d1-c03c460b5c77 req-aeeee0fd-3af1-4cbd-884d-4eb2d433101a service nova] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Received unexpected event network-vif-plugged-1bb5f5b0-d6c9-4b8d-be14-e6e98f2aeef1 for instance with vm_state building and task_state spawning. [ 960.208534] env[63531]: INFO nova.compute.resource_tracker [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Updating resource usage from migration d5d6b5dc-c0a9-41c2-96a9-6aa3ec666dcd [ 960.231884] env[63531]: DEBUG oslo_vmware.api [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1118271, 'name': MoveVirtualDisk_Task} progress is 57%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.479578] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e460709-4779-4f92-b342-f2db212e744e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.491187] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f6e1128-84b1-4583-b300-0b0b22b5c300 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.528984] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad9e9b72-5c3a-4d09-a3b2-07bbc195fe36 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.538390] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-078b0a85-a360-4067-8e21-ac6fdfe84555 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.554599] env[63531]: DEBUG nova.compute.provider_tree [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 960.592829] env[63531]: DEBUG oslo_vmware.api [None req-a384b8d7-53a1-4e7f-90b4-de586fca38e3 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118274, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.669898] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "refresh_cache-6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.670218] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquired lock "refresh_cache-6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.670325] env[63531]: DEBUG nova.network.neutron [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 960.729563] env[63531]: DEBUG oslo_vmware.api [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1118271, 'name': MoveVirtualDisk_Task} progress is 77%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.925433] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Acquiring lock "0fb3b687-62cc-4296-b4ed-537aba18c880" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.925761] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Lock "0fb3b687-62cc-4296-b4ed-537aba18c880" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.972169] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Acquiring lock "baf2db4e-2cdb-47f5-9a5c-97a233a459c9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.972169] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Lock "baf2db4e-2cdb-47f5-9a5c-97a233a459c9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.058013] env[63531]: DEBUG nova.scheduler.client.report [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 961.093985] env[63531]: DEBUG oslo_vmware.api [None req-a384b8d7-53a1-4e7f-90b4-de586fca38e3 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118274, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.228484] env[63531]: DEBUG oslo_vmware.api [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1118271, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.242792] env[63531]: DEBUG nova.network.neutron [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 961.428713] env[63531]: DEBUG nova.compute.manager [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: 0fb3b687-62cc-4296-b4ed-537aba18c880] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 961.475191] env[63531]: DEBUG nova.compute.manager [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: baf2db4e-2cdb-47f5-9a5c-97a233a459c9] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 961.490067] env[63531]: DEBUG nova.network.neutron [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Updating instance_info_cache with network_info: [{"id": "1bb5f5b0-d6c9-4b8d-be14-e6e98f2aeef1", "address": "fa:16:3e:30:5d:54", "network": {"id": "ea2d47d3-a502-4afd-9bc5-ee4a37017901", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1361282648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94a95667c00947ea9183a6307c569c90", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6c45fd-e930-495a-9cb7-df84eda443b1", "external-id": "nsx-vlan-transportzone-407", "segmentation_id": 407, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1bb5f5b0-d6", "ovs_interfaceid": "1bb5f5b0-d6c9-4b8d-be14-e6e98f2aeef1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 961.564159] env[63531]: DEBUG oslo_concurrency.lockutils [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.373s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.564649] env[63531]: INFO nova.compute.manager [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Migrating [ 961.574738] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "3bf08a09-b6e6-4a58-8d35-7061a8e59afd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.575019] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "3bf08a09-b6e6-4a58-8d35-7061a8e59afd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.578753] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a4bb48b8-dfee-4f65-99b6-ae50dabf2571 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.071s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.578989] env[63531]: DEBUG nova.objects.instance [None req-a4bb48b8-dfee-4f65-99b6-ae50dabf2571 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lazy-loading 'resources' on Instance uuid 48043b5d-c0d3-4e8a-8d28-5a408d800e11 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 961.596175] env[63531]: DEBUG oslo_vmware.api [None req-a384b8d7-53a1-4e7f-90b4-de586fca38e3 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118274, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.20774} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.596809] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-a384b8d7-53a1-4e7f-90b4-de586fca38e3 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 961.597008] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a384b8d7-53a1-4e7f-90b4-de586fca38e3 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 961.597230] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a384b8d7-53a1-4e7f-90b4-de586fca38e3 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 961.597362] env[63531]: INFO nova.compute.manager [None req-a384b8d7-53a1-4e7f-90b4-de586fca38e3 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Took 2.14 seconds to destroy the instance on the hypervisor. [ 961.597608] env[63531]: DEBUG oslo.service.loopingcall [None req-a384b8d7-53a1-4e7f-90b4-de586fca38e3 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 961.597828] env[63531]: DEBUG nova.compute.manager [-] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 961.598018] env[63531]: DEBUG nova.network.neutron [-] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 961.728222] env[63531]: DEBUG oslo_vmware.api [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1118271, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.593796} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.728508] env[63531]: INFO nova.virt.vmwareapi.ds_util [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_3e708714-2e59-4e1b-bd66-ad64677b1d1d/OSTACK_IMG_3e708714-2e59-4e1b-bd66-ad64677b1d1d.vmdk to [datastore1] devstack-image-cache_base/d2fca167-8767-4b16-b64a-61fd11e58a0e/d2fca167-8767-4b16-b64a-61fd11e58a0e.vmdk. [ 961.728710] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Cleaning up location [datastore1] OSTACK_IMG_3e708714-2e59-4e1b-bd66-ad64677b1d1d {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 961.728879] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_3e708714-2e59-4e1b-bd66-ad64677b1d1d {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 961.729140] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-55996652-4dc2-4419-b61e-da75bf965783 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.735852] env[63531]: DEBUG oslo_vmware.api [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Waiting for the task: (returnval){ [ 961.735852] env[63531]: value = "task-1118275" [ 961.735852] env[63531]: _type = "Task" [ 961.735852] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.743330] env[63531]: DEBUG oslo_vmware.api [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1118275, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.958239] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.983740] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Releasing lock "refresh_cache-6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 961.986025] env[63531]: DEBUG nova.compute.manager [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Instance network_info: |[{"id": "1bb5f5b0-d6c9-4b8d-be14-e6e98f2aeef1", "address": "fa:16:3e:30:5d:54", "network": {"id": "ea2d47d3-a502-4afd-9bc5-ee4a37017901", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1361282648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94a95667c00947ea9183a6307c569c90", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6c45fd-e930-495a-9cb7-df84eda443b1", "external-id": "nsx-vlan-transportzone-407", "segmentation_id": 407, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1bb5f5b0-d6", "ovs_interfaceid": "1bb5f5b0-d6c9-4b8d-be14-e6e98f2aeef1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 961.992559] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:30:5d:54', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2b6c45fd-e930-495a-9cb7-df84eda443b1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1bb5f5b0-d6c9-4b8d-be14-e6e98f2aeef1', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 962.003764] env[63531]: DEBUG oslo.service.loopingcall [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 962.003764] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 962.003764] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b04befb6-1397-4824-99da-f7b6d03bfb62 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.019572] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.025187] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 962.025187] env[63531]: value = "task-1118276" [ 962.025187] env[63531]: _type = "Task" [ 962.025187] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.032932] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118276, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.084594] env[63531]: DEBUG nova.compute.manager [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3bf08a09-b6e6-4a58-8d35-7061a8e59afd] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 962.092154] env[63531]: DEBUG oslo_concurrency.lockutils [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "refresh_cache-27c23b7c-a68f-43c9-a125-bd657feb3c5b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.092154] env[63531]: DEBUG oslo_concurrency.lockutils [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquired lock "refresh_cache-27c23b7c-a68f-43c9-a125-bd657feb3c5b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.092154] env[63531]: DEBUG nova.network.neutron [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 962.203710] env[63531]: DEBUG nova.compute.manager [req-569aab88-41e7-44e9-a030-0acb441c072c req-754b5efb-4100-49a6-a701-28b225b310d3 service nova] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Received event network-changed-1bb5f5b0-d6c9-4b8d-be14-e6e98f2aeef1 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 962.203805] env[63531]: DEBUG nova.compute.manager [req-569aab88-41e7-44e9-a030-0acb441c072c req-754b5efb-4100-49a6-a701-28b225b310d3 service nova] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Refreshing instance network info cache due to event network-changed-1bb5f5b0-d6c9-4b8d-be14-e6e98f2aeef1. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 962.204479] env[63531]: DEBUG oslo_concurrency.lockutils [req-569aab88-41e7-44e9-a030-0acb441c072c req-754b5efb-4100-49a6-a701-28b225b310d3 service nova] Acquiring lock "refresh_cache-6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.204729] env[63531]: DEBUG oslo_concurrency.lockutils [req-569aab88-41e7-44e9-a030-0acb441c072c req-754b5efb-4100-49a6-a701-28b225b310d3 service nova] Acquired lock "refresh_cache-6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.204905] env[63531]: DEBUG nova.network.neutron [req-569aab88-41e7-44e9-a030-0acb441c072c req-754b5efb-4100-49a6-a701-28b225b310d3 service nova] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Refreshing network info cache for port 1bb5f5b0-d6c9-4b8d-be14-e6e98f2aeef1 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 962.248545] env[63531]: DEBUG oslo_vmware.api [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1118275, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.033883} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.248880] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 962.249101] env[63531]: DEBUG oslo_concurrency.lockutils [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d2fca167-8767-4b16-b64a-61fd11e58a0e/d2fca167-8767-4b16-b64a-61fd11e58a0e.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.249383] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d2fca167-8767-4b16-b64a-61fd11e58a0e/d2fca167-8767-4b16-b64a-61fd11e58a0e.vmdk to [datastore1] 1ca7b851-2a06-4181-8271-58aafcd322d6/1ca7b851-2a06-4181-8271-58aafcd322d6.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 962.249643] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-82b356d9-a6f2-4ff5-9085-ad56db0b7fbe {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.256976] env[63531]: DEBUG oslo_vmware.api [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Waiting for the task: (returnval){ [ 962.256976] env[63531]: value = "task-1118277" [ 962.256976] env[63531]: _type = "Task" [ 962.256976] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.267242] env[63531]: DEBUG oslo_vmware.api [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1118277, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.382450] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0ae6423-2917-4c1d-8988-395bd56cf47e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.392147] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ff21a91-64bd-43bc-98cb-08512e3a9a1e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.425195] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22673b9b-6c52-461e-b107-cb339e92829d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.430700] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cec1578e-8d59-418c-867d-41d80ff508d3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.444428] env[63531]: DEBUG nova.compute.provider_tree [None req-a4bb48b8-dfee-4f65-99b6-ae50dabf2571 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 962.534538] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118276, 'name': CreateVM_Task, 'duration_secs': 0.407713} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.534715] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 962.535410] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.535588] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.535974] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 962.536231] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4255f572-0023-42f7-ba91-86bffbb8b932 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.540351] env[63531]: DEBUG oslo_vmware.api [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 962.540351] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52f1d62d-64ee-dcfb-2b10-45340812f9d8" [ 962.540351] env[63531]: _type = "Task" [ 962.540351] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.547238] env[63531]: DEBUG oslo_vmware.api [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52f1d62d-64ee-dcfb-2b10-45340812f9d8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.610132] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.652329] env[63531]: DEBUG nova.network.neutron [-] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.775015] env[63531]: DEBUG oslo_vmware.api [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1118277, 'name': CopyVirtualDisk_Task} progress is 12%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.854351] env[63531]: DEBUG nova.network.neutron [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Updating instance_info_cache with network_info: [{"id": "a755a88f-c86c-4183-b9fa-a6133060ea5e", "address": "fa:16:3e:26:e6:a5", "network": {"id": "bf99b8cb-9efb-4a74-9db1-0f5f0875f26a", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-998205595-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2e99349303e413b9f7896d449cb7dff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc6e6fe1-c4f5-4389-a49f-0978060eebb4", "external-id": "nsx-vlan-transportzone-829", "segmentation_id": 829, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa755a88f-c8", "ovs_interfaceid": "a755a88f-c86c-4183-b9fa-a6133060ea5e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.946942] env[63531]: DEBUG nova.scheduler.client.report [None req-a4bb48b8-dfee-4f65-99b6-ae50dabf2571 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 962.959180] env[63531]: DEBUG nova.network.neutron [req-569aab88-41e7-44e9-a030-0acb441c072c req-754b5efb-4100-49a6-a701-28b225b310d3 service nova] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Updated VIF entry in instance network info cache for port 1bb5f5b0-d6c9-4b8d-be14-e6e98f2aeef1. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 962.959877] env[63531]: DEBUG nova.network.neutron [req-569aab88-41e7-44e9-a030-0acb441c072c req-754b5efb-4100-49a6-a701-28b225b310d3 service nova] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Updating instance_info_cache with network_info: [{"id": "1bb5f5b0-d6c9-4b8d-be14-e6e98f2aeef1", "address": "fa:16:3e:30:5d:54", "network": {"id": "ea2d47d3-a502-4afd-9bc5-ee4a37017901", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1361282648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94a95667c00947ea9183a6307c569c90", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6c45fd-e930-495a-9cb7-df84eda443b1", "external-id": "nsx-vlan-transportzone-407", "segmentation_id": 407, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1bb5f5b0-d6", "ovs_interfaceid": "1bb5f5b0-d6c9-4b8d-be14-e6e98f2aeef1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.052465] env[63531]: DEBUG oslo_vmware.api [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52f1d62d-64ee-dcfb-2b10-45340812f9d8, 'name': SearchDatastore_Task, 'duration_secs': 0.066641} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.052856] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.053148] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 963.053419] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.053634] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.053869] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 963.054197] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5e6cf006-0a93-4b71-a1b2-ef54e11efecf {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.072613] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 963.072847] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 963.074326] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bf4c2815-32ab-44e1-a73c-860ca0a4c3cd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.082890] env[63531]: DEBUG oslo_vmware.api [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 963.082890] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52b565dc-2b3a-b678-64a6-75c1791d1f2b" [ 963.082890] env[63531]: _type = "Task" [ 963.082890] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.093204] env[63531]: DEBUG oslo_vmware.api [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52b565dc-2b3a-b678-64a6-75c1791d1f2b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.155470] env[63531]: INFO nova.compute.manager [-] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Took 1.56 seconds to deallocate network for instance. [ 963.273416] env[63531]: DEBUG oslo_vmware.api [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1118277, 'name': CopyVirtualDisk_Task} progress is 32%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.358203] env[63531]: DEBUG oslo_concurrency.lockutils [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Releasing lock "refresh_cache-27c23b7c-a68f-43c9-a125-bd657feb3c5b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.454100] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a4bb48b8-dfee-4f65-99b6-ae50dabf2571 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.875s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.456620] env[63531]: DEBUG oslo_concurrency.lockutils [None req-04a987cd-eb1f-407a-8135-fd56b4798069 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.828s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.456881] env[63531]: DEBUG nova.objects.instance [None req-04a987cd-eb1f-407a-8135-fd56b4798069 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Lazy-loading 'resources' on Instance uuid 55111b87-1e9a-4877-8689-987faa72a54e {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 963.463243] env[63531]: DEBUG oslo_concurrency.lockutils [req-569aab88-41e7-44e9-a030-0acb441c072c req-754b5efb-4100-49a6-a701-28b225b310d3 service nova] Releasing lock "refresh_cache-6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.463514] env[63531]: DEBUG nova.compute.manager [req-569aab88-41e7-44e9-a030-0acb441c072c req-754b5efb-4100-49a6-a701-28b225b310d3 service nova] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Received event network-vif-deleted-e4133e0f-55c1-4b3f-847d-899dd1167eec {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 963.463696] env[63531]: INFO nova.compute.manager [req-569aab88-41e7-44e9-a030-0acb441c072c req-754b5efb-4100-49a6-a701-28b225b310d3 service nova] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Neutron deleted interface e4133e0f-55c1-4b3f-847d-899dd1167eec; detaching it from the instance and deleting it from the info cache [ 963.464299] env[63531]: DEBUG nova.network.neutron [req-569aab88-41e7-44e9-a030-0acb441c072c req-754b5efb-4100-49a6-a701-28b225b310d3 service nova] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.479230] env[63531]: INFO nova.scheduler.client.report [None req-a4bb48b8-dfee-4f65-99b6-ae50dabf2571 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Deleted allocations for instance 48043b5d-c0d3-4e8a-8d28-5a408d800e11 [ 963.596284] env[63531]: DEBUG oslo_vmware.api [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52b565dc-2b3a-b678-64a6-75c1791d1f2b, 'name': SearchDatastore_Task, 'duration_secs': 0.074239} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.597166] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2ffcc1de-1e54-40b8-8561-98dfbbb4ab66 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.604893] env[63531]: DEBUG oslo_vmware.api [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 963.604893] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5249b649-673e-48c4-4b6b-995bfdfbb704" [ 963.604893] env[63531]: _type = "Task" [ 963.604893] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.615956] env[63531]: DEBUG oslo_vmware.api [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5249b649-673e-48c4-4b6b-995bfdfbb704, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.662770] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a384b8d7-53a1-4e7f-90b4-de586fca38e3 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.774343] env[63531]: DEBUG oslo_vmware.api [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1118277, 'name': CopyVirtualDisk_Task} progress is 52%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.967248] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-50f1b048-4dc1-4da5-a6dd-149b65153fdc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.980713] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccfa9fab-5d79-4060-b44d-760bc46bf755 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.005321] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a4bb48b8-dfee-4f65-99b6-ae50dabf2571 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "48043b5d-c0d3-4e8a-8d28-5a408d800e11" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.985s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.032474] env[63531]: DEBUG nova.compute.manager [req-569aab88-41e7-44e9-a030-0acb441c072c req-754b5efb-4100-49a6-a701-28b225b310d3 service nova] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Detach interface failed, port_id=e4133e0f-55c1-4b3f-847d-899dd1167eec, reason: Instance 9a3fad46-a15a-451c-bdab-a3c8cc8add07 could not be found. {{(pid=63531) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 964.121412] env[63531]: DEBUG oslo_vmware.api [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5249b649-673e-48c4-4b6b-995bfdfbb704, 'name': SearchDatastore_Task, 'duration_secs': 0.074939} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.121713] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 964.121992] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe/6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 964.122330] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1870f7d6-c7a1-4164-8046-2ac44227c3b2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.134415] env[63531]: DEBUG oslo_vmware.api [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 964.134415] env[63531]: value = "task-1118278" [ 964.134415] env[63531]: _type = "Task" [ 964.134415] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.148047] env[63531]: DEBUG oslo_vmware.api [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118278, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.256434] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84218d91-ce34-418b-b57c-e9a9cc6c90d7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.269380] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33e25d9d-26bd-49cc-bc63-dff654560f18 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.276183] env[63531]: DEBUG oslo_vmware.api [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1118277, 'name': CopyVirtualDisk_Task} progress is 74%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.302212] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7eb8fca1-3471-4dbd-9188-b9cd65a832ba {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.311619] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0ce5c49-6487-4571-a2cd-53f94f095910 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.328020] env[63531]: DEBUG nova.compute.provider_tree [None req-04a987cd-eb1f-407a-8135-fd56b4798069 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 964.647038] env[63531]: DEBUG oslo_vmware.api [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118278, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.774783] env[63531]: DEBUG oslo_vmware.api [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1118277, 'name': CopyVirtualDisk_Task} progress is 94%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.831575] env[63531]: DEBUG nova.scheduler.client.report [None req-04a987cd-eb1f-407a-8135-fd56b4798069 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 964.874323] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77afc978-c443-4451-bd81-cd48cf19ec9f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.893238] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Updating instance '27c23b7c-a68f-43c9-a125-bd657feb3c5b' progress to 0 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 965.147127] env[63531]: DEBUG oslo_vmware.api [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118278, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.273522] env[63531]: DEBUG oslo_vmware.api [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1118277, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.627894} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.273688] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d2fca167-8767-4b16-b64a-61fd11e58a0e/d2fca167-8767-4b16-b64a-61fd11e58a0e.vmdk to [datastore1] 1ca7b851-2a06-4181-8271-58aafcd322d6/1ca7b851-2a06-4181-8271-58aafcd322d6.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 965.274464] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db535e6b-0b1a-4d4c-a685-5d2c984ca390 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.295975] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Reconfiguring VM instance instance-0000002a to attach disk [datastore1] 1ca7b851-2a06-4181-8271-58aafcd322d6/1ca7b851-2a06-4181-8271-58aafcd322d6.vmdk or device None with type streamOptimized {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 965.296383] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-67f0454f-b53e-4938-89b6-6bf867c9288e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.315568] env[63531]: DEBUG oslo_vmware.api [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Waiting for the task: (returnval){ [ 965.315568] env[63531]: value = "task-1118279" [ 965.315568] env[63531]: _type = "Task" [ 965.315568] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.322871] env[63531]: DEBUG oslo_vmware.api [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1118279, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.336770] env[63531]: DEBUG oslo_concurrency.lockutils [None req-04a987cd-eb1f-407a-8135-fd56b4798069 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.880s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.339250] env[63531]: DEBUG oslo_concurrency.lockutils [None req-293635f3-1f0c-487a-ad12-e4a71b80c8dd tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.260s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.339516] env[63531]: DEBUG nova.objects.instance [None req-293635f3-1f0c-487a-ad12-e4a71b80c8dd tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lazy-loading 'resources' on Instance uuid a090135f-489d-457d-be36-ba6f61b71ab8 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 965.362754] env[63531]: INFO nova.scheduler.client.report [None req-04a987cd-eb1f-407a-8135-fd56b4798069 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Deleted allocations for instance 55111b87-1e9a-4877-8689-987faa72a54e [ 965.398890] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 965.399218] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1092a44d-9bb8-4ac7-a343-76db5420c498 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.406325] env[63531]: DEBUG oslo_vmware.api [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 965.406325] env[63531]: value = "task-1118280" [ 965.406325] env[63531]: _type = "Task" [ 965.406325] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.415327] env[63531]: DEBUG oslo_vmware.api [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118280, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.646512] env[63531]: DEBUG oslo_vmware.api [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118278, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.118657} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.646798] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe/6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 965.647048] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 965.647297] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8d87e8c9-703b-42d1-b8e1-576b228ea2df {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.653321] env[63531]: DEBUG oslo_vmware.api [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 965.653321] env[63531]: value = "task-1118281" [ 965.653321] env[63531]: _type = "Task" [ 965.653321] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.664333] env[63531]: DEBUG oslo_vmware.api [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118281, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.826336] env[63531]: DEBUG oslo_vmware.api [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1118279, 'name': ReconfigVM_Task, 'duration_secs': 0.280238} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.826633] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Reconfigured VM instance instance-0000002a to attach disk [datastore1] 1ca7b851-2a06-4181-8271-58aafcd322d6/1ca7b851-2a06-4181-8271-58aafcd322d6.vmdk or device None with type streamOptimized {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 965.827389] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a1096cb6-0eaa-44c5-82da-ea513e3c27a4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.833449] env[63531]: DEBUG oslo_vmware.api [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Waiting for the task: (returnval){ [ 965.833449] env[63531]: value = "task-1118282" [ 965.833449] env[63531]: _type = "Task" [ 965.833449] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.841237] env[63531]: DEBUG oslo_vmware.api [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1118282, 'name': Rename_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.869059] env[63531]: DEBUG oslo_concurrency.lockutils [None req-04a987cd-eb1f-407a-8135-fd56b4798069 tempest-ServerPasswordTestJSON-902330126 tempest-ServerPasswordTestJSON-902330126-project-member] Lock "55111b87-1e9a-4877-8689-987faa72a54e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.228s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.917156] env[63531]: DEBUG oslo_vmware.api [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118280, 'name': PowerOffVM_Task, 'duration_secs': 0.193088} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.920383] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 965.920383] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Updating instance '27c23b7c-a68f-43c9-a125-bd657feb3c5b' progress to 17 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 966.095555] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ccb1a3f-09b1-4cb9-a2c1-0c2d3c6084dc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.103231] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae72a141-ef10-4cd4-9255-f36574e6295b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.133826] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0eae458-ee8e-48e0-9039-6dd13618556d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.140860] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c517a2d9-b240-47bb-83cc-bb1c2e05d36f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.153645] env[63531]: DEBUG nova.compute.provider_tree [None req-293635f3-1f0c-487a-ad12-e4a71b80c8dd tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 966.161949] env[63531]: DEBUG oslo_vmware.api [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118281, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.266744} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.162783] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 966.163537] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e21f8cc2-e423-49a9-acb4-c64ba89a451c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.186711] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Reconfiguring VM instance instance-00000056 to attach disk [datastore1] 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe/6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 966.187206] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-666d26c7-1c2b-41aa-a240-81b9e80fa7bc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.205454] env[63531]: DEBUG oslo_vmware.api [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 966.205454] env[63531]: value = "task-1118283" [ 966.205454] env[63531]: _type = "Task" [ 966.205454] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.212896] env[63531]: DEBUG oslo_vmware.api [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118283, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.343134] env[63531]: DEBUG oslo_vmware.api [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1118282, 'name': Rename_Task, 'duration_secs': 0.181542} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.343511] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 966.343821] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c44ac7d3-8adf-4b91-ac06-66bed780f562 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.350431] env[63531]: DEBUG oslo_vmware.api [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Waiting for the task: (returnval){ [ 966.350431] env[63531]: value = "task-1118284" [ 966.350431] env[63531]: _type = "Task" [ 966.350431] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.357972] env[63531]: DEBUG oslo_vmware.api [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1118284, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.429185] env[63531]: DEBUG nova.virt.hardware [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 966.429185] env[63531]: DEBUG nova.virt.hardware [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 966.429185] env[63531]: DEBUG nova.virt.hardware [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 966.429185] env[63531]: DEBUG nova.virt.hardware [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 966.429695] env[63531]: DEBUG nova.virt.hardware [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 966.429992] env[63531]: DEBUG nova.virt.hardware [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 966.430352] env[63531]: DEBUG nova.virt.hardware [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 966.430649] env[63531]: DEBUG nova.virt.hardware [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 966.430973] env[63531]: DEBUG nova.virt.hardware [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 966.431356] env[63531]: DEBUG nova.virt.hardware [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 966.431708] env[63531]: DEBUG nova.virt.hardware [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 966.438166] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b23562ef-d97d-4bc6-961f-2fb5dfbbd2f2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.456328] env[63531]: DEBUG oslo_vmware.api [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 966.456328] env[63531]: value = "task-1118285" [ 966.456328] env[63531]: _type = "Task" [ 966.456328] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.464386] env[63531]: DEBUG oslo_vmware.api [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118285, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.594415] env[63531]: DEBUG oslo_concurrency.lockutils [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquiring lock "4757625a-8398-4b2d-b6fe-40eea8913068" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.594415] env[63531]: DEBUG oslo_concurrency.lockutils [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "4757625a-8398-4b2d-b6fe-40eea8913068" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.659846] env[63531]: DEBUG nova.scheduler.client.report [None req-293635f3-1f0c-487a-ad12-e4a71b80c8dd tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 966.719026] env[63531]: DEBUG oslo_vmware.api [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118283, 'name': ReconfigVM_Task, 'duration_secs': 0.380103} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.719775] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Reconfigured VM instance instance-00000056 to attach disk [datastore1] 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe/6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 966.720264] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6beec6c9-4f85-4135-9344-0508c1c3b528 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.731880] env[63531]: DEBUG oslo_vmware.api [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 966.731880] env[63531]: value = "task-1118286" [ 966.731880] env[63531]: _type = "Task" [ 966.731880] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.743988] env[63531]: DEBUG oslo_vmware.api [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118286, 'name': Rename_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.861519] env[63531]: DEBUG oslo_vmware.api [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1118284, 'name': PowerOnVM_Task, 'duration_secs': 0.467942} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.861803] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 966.965983] env[63531]: DEBUG oslo_vmware.api [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118285, 'name': ReconfigVM_Task, 'duration_secs': 0.288759} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.966806] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Updating instance '27c23b7c-a68f-43c9-a125-bd657feb3c5b' progress to 33 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 966.971126] env[63531]: DEBUG nova.compute.manager [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 966.974342] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a018ac34-ea07-4fa4-9483-6b21cf427280 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.096796] env[63531]: DEBUG nova.compute.manager [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 967.165883] env[63531]: DEBUG oslo_concurrency.lockutils [None req-293635f3-1f0c-487a-ad12-e4a71b80c8dd tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.827s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.168591] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 10.830s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.196622] env[63531]: INFO nova.scheduler.client.report [None req-293635f3-1f0c-487a-ad12-e4a71b80c8dd tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Deleted allocations for instance a090135f-489d-457d-be36-ba6f61b71ab8 [ 967.242657] env[63531]: DEBUG oslo_vmware.api [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118286, 'name': Rename_Task, 'duration_secs': 0.138295} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.242961] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 967.243234] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9f6c853d-ca59-42f2-8cf4-ebe7694dd7e2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.249715] env[63531]: DEBUG oslo_vmware.api [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 967.249715] env[63531]: value = "task-1118287" [ 967.249715] env[63531]: _type = "Task" [ 967.249715] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.257239] env[63531]: DEBUG oslo_vmware.api [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118287, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.479668] env[63531]: DEBUG nova.virt.hardware [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 967.480097] env[63531]: DEBUG nova.virt.hardware [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 967.480371] env[63531]: DEBUG nova.virt.hardware [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 967.480700] env[63531]: DEBUG nova.virt.hardware [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 967.480934] env[63531]: DEBUG nova.virt.hardware [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 967.482185] env[63531]: DEBUG nova.virt.hardware [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 967.482185] env[63531]: DEBUG nova.virt.hardware [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 967.482185] env[63531]: DEBUG nova.virt.hardware [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 967.482185] env[63531]: DEBUG nova.virt.hardware [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 967.482185] env[63531]: DEBUG nova.virt.hardware [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 967.482185] env[63531]: DEBUG nova.virt.hardware [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 967.487524] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Reconfiguring VM instance instance-00000053 to detach disk 2000 {{(pid=63531) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 967.491524] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0efdec27-bb5e-466a-bd20-f2ec716513aa {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.507601] env[63531]: DEBUG oslo_concurrency.lockutils [None req-729d0c85-b692-414e-9580-a6805080dc40 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Lock "1ca7b851-2a06-4181-8271-58aafcd322d6" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 39.516s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.513699] env[63531]: DEBUG oslo_vmware.api [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 967.513699] env[63531]: value = "task-1118288" [ 967.513699] env[63531]: _type = "Task" [ 967.513699] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.522222] env[63531]: DEBUG oslo_vmware.api [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118288, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.615641] env[63531]: DEBUG oslo_concurrency.lockutils [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.673087] env[63531]: INFO nova.compute.claims [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 967.705293] env[63531]: DEBUG oslo_concurrency.lockutils [None req-293635f3-1f0c-487a-ad12-e4a71b80c8dd tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "a090135f-489d-457d-be36-ba6f61b71ab8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.062s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.706453] env[63531]: DEBUG oslo_concurrency.lockutils [req-2dbf38a6-5e61-4daf-afb5-6104396969e1 req-bc7e5bb5-df3f-4730-94ed-e4e09f0e4e8c service nova] Acquired lock "a090135f-489d-457d-be36-ba6f61b71ab8" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.707829] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f682339c-5d25-4f68-9b46-a39a4deca56b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.716095] env[63531]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 967.716263] env[63531]: DEBUG oslo_vmware.api [-] Fault list: [ManagedObjectNotFound] {{(pid=63531) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 967.716702] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5d093d84-b5b2-4cb8-b4ff-f9909dba1f7a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.725907] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21f8afda-ac93-4857-9799-114b66502ce1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.756930] env[63531]: ERROR root [req-2dbf38a6-5e61-4daf-afb5-6104396969e1 req-bc7e5bb5-df3f-4730-94ed-e4e09f0e4e8c service nova] Original exception being dropped: ['Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 377, in request_handler\n response = request(managed_object, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 586, in __call__\n return client.invoke(args, kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 728, in invoke\n result = self.send(soapenv, timeout=timeout)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 777, in send\n return self.process_reply(reply.message, None, None)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 840, in process_reply\n raise WebFault(fault, replyroot)\n', "suds.WebFault: Server raised fault: 'The object 'vim.VirtualMachine:vm-244734' has already been deleted or has not been completely created'\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 301, in _invoke_api\n return api_method(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 480, in get_object_property\n props = get_object_properties(vim, moref, [property_name],\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 360, in get_object_properties\n retrieve_result = vim.RetrievePropertiesEx(\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 413, in request_handler\n raise exceptions.VimFaultException(fault_list, fault_string,\n', "oslo_vmware.exceptions.VimFaultException: The object 'vim.VirtualMachine:vm-244734' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-244734' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-244734'}\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/session.py", line 123, in _call_method\n return self.invoke_api(module, method, self.vim, *args,\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 358, in invoke_api\n return _invoke_api(module, method, *args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 122, in func\n return evt.wait()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait\n result = hub.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch\n return self.greenlet.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/common/loopingcall.py", line 122, in _inner\n idle = self.f(*self.args, **self.kw)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 96, in _func\n result = f(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 341, in _invoke_api\n raise clazz(str(excep),\n', "oslo_vmware.exceptions.ManagedObjectNotFoundException: The object 'vim.VirtualMachine:vm-244734' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-244734' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-244734'}\n"]: nova.exception.InstanceNotFound: Instance a090135f-489d-457d-be36-ba6f61b71ab8 could not be found. [ 967.757161] env[63531]: DEBUG oslo_concurrency.lockutils [req-2dbf38a6-5e61-4daf-afb5-6104396969e1 req-bc7e5bb5-df3f-4730-94ed-e4e09f0e4e8c service nova] Releasing lock "a090135f-489d-457d-be36-ba6f61b71ab8" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.757386] env[63531]: DEBUG nova.compute.manager [req-2dbf38a6-5e61-4daf-afb5-6104396969e1 req-bc7e5bb5-df3f-4730-94ed-e4e09f0e4e8c service nova] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Detach interface failed, port_id=1d5fea6d-10c4-4740-81a1-8e1ac927dad5, reason: Instance a090135f-489d-457d-be36-ba6f61b71ab8 could not be found. {{(pid=63531) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 967.768982] env[63531]: DEBUG oslo_vmware.api [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118287, 'name': PowerOnVM_Task, 'duration_secs': 0.453924} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.769286] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 967.769486] env[63531]: INFO nova.compute.manager [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Took 9.13 seconds to spawn the instance on the hypervisor. [ 967.769667] env[63531]: DEBUG nova.compute.manager [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 967.770994] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0c15aa5-d9af-4ab9-af89-8f556767e386 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.023934] env[63531]: DEBUG oslo_vmware.api [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118288, 'name': ReconfigVM_Task, 'duration_secs': 0.165319} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.024263] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Reconfigured VM instance instance-00000053 to detach disk 2000 {{(pid=63531) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 968.025058] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c25e9f7-f60e-443d-a23e-c41f37087678 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.047678] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Reconfiguring VM instance instance-00000053 to attach disk [datastore1] 27c23b7c-a68f-43c9-a125-bd657feb3c5b/27c23b7c-a68f-43c9-a125-bd657feb3c5b.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 968.048017] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e5567ad5-7fdc-497c-acdf-aa22134b3540 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.066641] env[63531]: DEBUG oslo_vmware.api [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 968.066641] env[63531]: value = "task-1118289" [ 968.066641] env[63531]: _type = "Task" [ 968.066641] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.077975] env[63531]: DEBUG oslo_vmware.api [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118289, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.179708] env[63531]: INFO nova.compute.resource_tracker [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Updating resource usage from migration 731cf9af-2cef-4013-8292-6c5aedf6f9ee [ 968.246604] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9609debe-f9ca-4390-a57b-86b7f7ea872e tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Acquiring lock "f46d8a85-6ab7-436a-bc5e-40d0f80974e5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.246960] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9609debe-f9ca-4390-a57b-86b7f7ea872e tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lock "f46d8a85-6ab7-436a-bc5e-40d0f80974e5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.247252] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9609debe-f9ca-4390-a57b-86b7f7ea872e tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Acquiring lock "f46d8a85-6ab7-436a-bc5e-40d0f80974e5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.247437] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9609debe-f9ca-4390-a57b-86b7f7ea872e tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lock "f46d8a85-6ab7-436a-bc5e-40d0f80974e5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.247867] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9609debe-f9ca-4390-a57b-86b7f7ea872e tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lock "f46d8a85-6ab7-436a-bc5e-40d0f80974e5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.250111] env[63531]: INFO nova.compute.manager [None req-9609debe-f9ca-4390-a57b-86b7f7ea872e tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Terminating instance [ 968.254562] env[63531]: DEBUG nova.compute.manager [None req-9609debe-f9ca-4390-a57b-86b7f7ea872e tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 968.254845] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-9609debe-f9ca-4390-a57b-86b7f7ea872e tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 968.256143] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14957b02-3e2a-426c-ac65-e31e9a4c5d25 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.265082] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-9609debe-f9ca-4390-a57b-86b7f7ea872e tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 968.265775] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2138b1e4-c6cf-46d2-90f0-2b9cefb7226c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.280114] env[63531]: DEBUG oslo_vmware.api [None req-9609debe-f9ca-4390-a57b-86b7f7ea872e tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Waiting for the task: (returnval){ [ 968.280114] env[63531]: value = "task-1118290" [ 968.280114] env[63531]: _type = "Task" [ 968.280114] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.301263] env[63531]: DEBUG oslo_vmware.api [None req-9609debe-f9ca-4390-a57b-86b7f7ea872e tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118290, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.301913] env[63531]: INFO nova.compute.manager [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Took 26.01 seconds to build instance. [ 968.469286] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2484bfd4-e379-48f0-9df5-15c73e9e809f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.477121] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ac48d27-e845-43aa-90c4-e5da1d723de2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.509991] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e66afd1-e76b-4ae5-8941-be88decbf50e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.517571] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f757033-5258-449c-8f88-1362e4f1bfdf {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.531814] env[63531]: DEBUG nova.compute.provider_tree [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 968.578806] env[63531]: DEBUG oslo_vmware.api [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118289, 'name': ReconfigVM_Task, 'duration_secs': 0.287303} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.579965] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Reconfigured VM instance instance-00000053 to attach disk [datastore1] 27c23b7c-a68f-43c9-a125-bd657feb3c5b/27c23b7c-a68f-43c9-a125-bd657feb3c5b.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 968.579965] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Updating instance '27c23b7c-a68f-43c9-a125-bd657feb3c5b' progress to 50 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 968.790689] env[63531]: DEBUG oslo_vmware.api [None req-9609debe-f9ca-4390-a57b-86b7f7ea872e tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118290, 'name': PowerOffVM_Task, 'duration_secs': 0.382961} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.791481] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-9609debe-f9ca-4390-a57b-86b7f7ea872e tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 968.791603] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-9609debe-f9ca-4390-a57b-86b7f7ea872e tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 968.791803] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-52a5a2c6-9370-42b6-804c-778e09e688ba {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.804159] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62018266-b0af-472c-8e62-f664da768017 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.528s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.855444] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-9609debe-f9ca-4390-a57b-86b7f7ea872e tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 968.855637] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-9609debe-f9ca-4390-a57b-86b7f7ea872e tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Deleting contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 968.855856] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-9609debe-f9ca-4390-a57b-86b7f7ea872e tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Deleting the datastore file [datastore2] f46d8a85-6ab7-436a-bc5e-40d0f80974e5 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 968.856254] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-51f7fde3-e8f0-4815-9072-08ca0ed3eb0c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.862814] env[63531]: DEBUG oslo_vmware.api [None req-9609debe-f9ca-4390-a57b-86b7f7ea872e tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Waiting for the task: (returnval){ [ 968.862814] env[63531]: value = "task-1118292" [ 968.862814] env[63531]: _type = "Task" [ 968.862814] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.871111] env[63531]: DEBUG oslo_vmware.api [None req-9609debe-f9ca-4390-a57b-86b7f7ea872e tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118292, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.034947] env[63531]: DEBUG nova.scheduler.client.report [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 969.091453] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7215a937-d333-41d5-8a98-b5332af46d73 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.119354] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc633f43-f482-4430-b1a1-822404b25800 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.137830] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Updating instance '27c23b7c-a68f-43c9-a125-bd657feb3c5b' progress to 67 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 969.373301] env[63531]: DEBUG oslo_vmware.api [None req-9609debe-f9ca-4390-a57b-86b7f7ea872e tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118292, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141067} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.373584] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-9609debe-f9ca-4390-a57b-86b7f7ea872e tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 969.373774] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-9609debe-f9ca-4390-a57b-86b7f7ea872e tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Deleted contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 969.374011] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-9609debe-f9ca-4390-a57b-86b7f7ea872e tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 969.374211] env[63531]: INFO nova.compute.manager [None req-9609debe-f9ca-4390-a57b-86b7f7ea872e tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Took 1.12 seconds to destroy the instance on the hypervisor. [ 969.374459] env[63531]: DEBUG oslo.service.loopingcall [None req-9609debe-f9ca-4390-a57b-86b7f7ea872e tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 969.374658] env[63531]: DEBUG nova.compute.manager [-] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 969.374754] env[63531]: DEBUG nova.network.neutron [-] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 969.542030] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.372s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.542030] env[63531]: INFO nova.compute.manager [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Migrating [ 969.548501] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.590s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.550640] env[63531]: INFO nova.compute.claims [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: 0fb3b687-62cc-4296-b4ed-537aba18c880] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 969.679928] env[63531]: DEBUG nova.network.neutron [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Port a755a88f-c86c-4183-b9fa-a6133060ea5e binding to destination host cpu-1 is already ACTIVE {{(pid=63531) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 969.743917] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-730d087d-72f9-4dec-a1f5-d292017cb4bd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.751897] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f964b1c1-ba3d-4569-a668-41df941c710a tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Suspending the VM {{(pid=63531) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 969.751897] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-63c033b5-3ff0-42eb-879f-399367f8cdf3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.758544] env[63531]: DEBUG oslo_vmware.api [None req-f964b1c1-ba3d-4569-a668-41df941c710a tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Waiting for the task: (returnval){ [ 969.758544] env[63531]: value = "task-1118293" [ 969.758544] env[63531]: _type = "Task" [ 969.758544] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.768243] env[63531]: DEBUG oslo_vmware.api [None req-f964b1c1-ba3d-4569-a668-41df941c710a tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1118293, 'name': SuspendVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.066161] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "refresh_cache-29dd6de3-2765-408c-acc0-da47e5e0a977" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 970.066351] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquired lock "refresh_cache-29dd6de3-2765-408c-acc0-da47e5e0a977" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.066584] env[63531]: DEBUG nova.network.neutron [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 970.121052] env[63531]: DEBUG nova.compute.manager [req-955198ea-de6b-49fc-b23b-3086e1f2feaf req-f7968871-5575-4a39-b64c-16acd8cc010c service nova] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Received event network-vif-deleted-a3fd640b-c47b-4501-ac04-f90f6af8b076 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 970.121329] env[63531]: INFO nova.compute.manager [req-955198ea-de6b-49fc-b23b-3086e1f2feaf req-f7968871-5575-4a39-b64c-16acd8cc010c service nova] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Neutron deleted interface a3fd640b-c47b-4501-ac04-f90f6af8b076; detaching it from the instance and deleting it from the info cache [ 970.121437] env[63531]: DEBUG nova.network.neutron [req-955198ea-de6b-49fc-b23b-3086e1f2feaf req-f7968871-5575-4a39-b64c-16acd8cc010c service nova] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 970.274568] env[63531]: DEBUG oslo_vmware.api [None req-f964b1c1-ba3d-4569-a668-41df941c710a tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1118293, 'name': SuspendVM_Task} progress is 50%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.433998] env[63531]: DEBUG oslo_concurrency.lockutils [None req-d93e69f3-6cef-48fe-bf5f-c50010578386 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.434339] env[63531]: DEBUG oslo_concurrency.lockutils [None req-d93e69f3-6cef-48fe-bf5f-c50010578386 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.484603] env[63531]: DEBUG nova.network.neutron [-] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 970.624665] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0166763b-43d3-4184-8421-019546d316f2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.634582] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45aaddbb-39d0-4710-ba50-1b0d88d23d43 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.671279] env[63531]: DEBUG nova.compute.manager [req-955198ea-de6b-49fc-b23b-3086e1f2feaf req-f7968871-5575-4a39-b64c-16acd8cc010c service nova] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Detach interface failed, port_id=a3fd640b-c47b-4501-ac04-f90f6af8b076, reason: Instance f46d8a85-6ab7-436a-bc5e-40d0f80974e5 could not be found. {{(pid=63531) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 970.706907] env[63531]: DEBUG oslo_concurrency.lockutils [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "27c23b7c-a68f-43c9-a125-bd657feb3c5b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.706907] env[63531]: DEBUG oslo_concurrency.lockutils [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "27c23b7c-a68f-43c9-a125-bd657feb3c5b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.706907] env[63531]: DEBUG oslo_concurrency.lockutils [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "27c23b7c-a68f-43c9-a125-bd657feb3c5b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.768321] env[63531]: DEBUG oslo_vmware.api [None req-f964b1c1-ba3d-4569-a668-41df941c710a tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1118293, 'name': SuspendVM_Task, 'duration_secs': 0.955391} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.771121] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f964b1c1-ba3d-4569-a668-41df941c710a tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Suspended the VM {{(pid=63531) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 970.771356] env[63531]: DEBUG nova.compute.manager [None req-f964b1c1-ba3d-4569-a668-41df941c710a tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 970.772384] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21a366c5-b9b9-4cdd-b82b-a76db4a6ea97 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.827958] env[63531]: DEBUG nova.network.neutron [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Updating instance_info_cache with network_info: [{"id": "262abfb7-7bcf-40e6-9ce3-b502de45b3b4", "address": "fa:16:3e:0e:00:c9", "network": {"id": "0667c176-cfdf-4e43-b152-e00bdc224448", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-243715985-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.217", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "308eabafea9c4188a58a0f1c22074d2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8f580e6-1d86-41ee-9ebe-c531cb9299c6", "external-id": "nsx-vlan-transportzone-150", "segmentation_id": 150, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap262abfb7-7b", "ovs_interfaceid": "262abfb7-7bcf-40e6-9ce3-b502de45b3b4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 970.882574] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "e742d3af-5a45-4274-adfb-c6138bf84d6a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.882833] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "e742d3af-5a45-4274-adfb-c6138bf84d6a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.896478] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9946d15a-0b46-4fff-a8af-c1e91382d625 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.909820] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aff07f6-f468-4617-9c97-6a37394bc0d8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.648564] env[63531]: DEBUG nova.compute.utils [None req-d93e69f3-6cef-48fe-bf5f-c50010578386 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 971.650028] env[63531]: INFO nova.compute.manager [-] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Took 2.28 seconds to deallocate network for instance. [ 971.653361] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Releasing lock "refresh_cache-29dd6de3-2765-408c-acc0-da47e5e0a977" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 971.654516] env[63531]: DEBUG nova.compute.manager [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 971.693345] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f30b76e-8808-4209-97d5-ef053d7d11d9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.704149] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-923404ee-e47e-4077-b243-d24a5417d64d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.720008] env[63531]: DEBUG nova.compute.provider_tree [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 972.158912] env[63531]: DEBUG oslo_concurrency.lockutils [None req-d93e69f3-6cef-48fe-bf5f-c50010578386 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.724s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.193814] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9609debe-f9ca-4390-a57b-86b7f7ea872e tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.195096] env[63531]: DEBUG oslo_concurrency.lockutils [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "refresh_cache-27c23b7c-a68f-43c9-a125-bd657feb3c5b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 972.195274] env[63531]: DEBUG oslo_concurrency.lockutils [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquired lock "refresh_cache-27c23b7c-a68f-43c9-a125-bd657feb3c5b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.195451] env[63531]: DEBUG nova.network.neutron [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 972.223045] env[63531]: DEBUG nova.scheduler.client.report [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 972.230305] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.434358] env[63531]: INFO nova.compute.manager [None req-16b6231f-55ef-4d9c-ab7e-83b12b5da861 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Resuming [ 972.435205] env[63531]: DEBUG nova.objects.instance [None req-16b6231f-55ef-4d9c-ab7e-83b12b5da861 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Lazy-loading 'flavor' on Instance uuid 1ca7b851-2a06-4181-8271-58aafcd322d6 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 972.728496] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.180s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.729047] env[63531]: DEBUG nova.compute.manager [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: 0fb3b687-62cc-4296-b4ed-537aba18c880] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 972.731667] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.712s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.733010] env[63531]: INFO nova.compute.claims [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: baf2db4e-2cdb-47f5-9a5c-97a233a459c9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 972.904081] env[63531]: DEBUG nova.network.neutron [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Updating instance_info_cache with network_info: [{"id": "a755a88f-c86c-4183-b9fa-a6133060ea5e", "address": "fa:16:3e:26:e6:a5", "network": {"id": "bf99b8cb-9efb-4a74-9db1-0f5f0875f26a", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-998205595-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2e99349303e413b9f7896d449cb7dff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc6e6fe1-c4f5-4389-a49f-0978060eebb4", "external-id": "nsx-vlan-transportzone-829", "segmentation_id": 829, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa755a88f-c8", "ovs_interfaceid": "a755a88f-c86c-4183-b9fa-a6133060ea5e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 973.007218] env[63531]: DEBUG oslo_concurrency.lockutils [None req-d93e69f3-6cef-48fe-bf5f-c50010578386 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.007490] env[63531]: DEBUG oslo_concurrency.lockutils [None req-d93e69f3-6cef-48fe-bf5f-c50010578386 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.007772] env[63531]: INFO nova.compute.manager [None req-d93e69f3-6cef-48fe-bf5f-c50010578386 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Attaching volume 332ac07f-9532-470d-aae2-732a5fea0e2a to /dev/sdb [ 973.041535] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aadd755a-2d5e-4f4d-8348-6990c8d4db9e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.048657] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76d36638-974f-4e8c-8caf-e5c5d02ce70b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.061940] env[63531]: DEBUG nova.virt.block_device [None req-d93e69f3-6cef-48fe-bf5f-c50010578386 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Updating existing volume attachment record: df97ed12-7e43-405d-a0da-d72d1a7f698f {{(pid=63531) _volume_attach /opt/stack/nova/nova/virt/block_device.py:679}} [ 973.176240] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f043bc9c-2a8a-442d-9feb-cf8f6dc75f66 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.195086] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Updating instance '29dd6de3-2765-408c-acc0-da47e5e0a977' progress to 0 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 973.240161] env[63531]: DEBUG nova.compute.utils [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 973.243111] env[63531]: DEBUG nova.compute.manager [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: 0fb3b687-62cc-4296-b4ed-537aba18c880] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 973.243299] env[63531]: DEBUG nova.network.neutron [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: 0fb3b687-62cc-4296-b4ed-537aba18c880] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 973.283635] env[63531]: DEBUG nova.policy [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4186052665ce457b99e1fee971be05b4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4ab5ee56531c4e638f8b695b9e5d082f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 973.407348] env[63531]: DEBUG oslo_concurrency.lockutils [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Releasing lock "refresh_cache-27c23b7c-a68f-43c9-a125-bd657feb3c5b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.447026] env[63531]: DEBUG oslo_concurrency.lockutils [None req-16b6231f-55ef-4d9c-ab7e-83b12b5da861 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Acquiring lock "refresh_cache-1ca7b851-2a06-4181-8271-58aafcd322d6" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 973.447026] env[63531]: DEBUG oslo_concurrency.lockutils [None req-16b6231f-55ef-4d9c-ab7e-83b12b5da861 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Acquired lock "refresh_cache-1ca7b851-2a06-4181-8271-58aafcd322d6" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 973.447026] env[63531]: DEBUG nova.network.neutron [None req-16b6231f-55ef-4d9c-ab7e-83b12b5da861 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 973.701504] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 973.701885] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bdc6dbc9-78a8-4ec7-bafa-17c74ec78a92 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.712511] env[63531]: DEBUG oslo_vmware.api [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 973.712511] env[63531]: value = "task-1118297" [ 973.712511] env[63531]: _type = "Task" [ 973.712511] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.720955] env[63531]: DEBUG oslo_vmware.api [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118297, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.746716] env[63531]: DEBUG nova.network.neutron [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: 0fb3b687-62cc-4296-b4ed-537aba18c880] Successfully created port: ac505e2f-66a8-467d-b109-d86556eedc79 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 973.748867] env[63531]: DEBUG nova.compute.manager [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: 0fb3b687-62cc-4296-b4ed-537aba18c880] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 973.753970] env[63531]: DEBUG oslo_concurrency.lockutils [None req-fc9401a8-2566-43ef-b491-055728bc20c4 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquiring lock "ccf00b09-29fe-4a76-a8af-97774acd77de" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.754645] env[63531]: DEBUG oslo_concurrency.lockutils [None req-fc9401a8-2566-43ef-b491-055728bc20c4 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Lock "ccf00b09-29fe-4a76-a8af-97774acd77de" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.851942] env[63531]: DEBUG oslo_concurrency.lockutils [None req-795646f8-c711-4669-82d5-9d0c4ed80a2a tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Acquiring lock "2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.851942] env[63531]: DEBUG oslo_concurrency.lockutils [None req-795646f8-c711-4669-82d5-9d0c4ed80a2a tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Lock "2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.851942] env[63531]: DEBUG oslo_concurrency.lockutils [None req-795646f8-c711-4669-82d5-9d0c4ed80a2a tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Acquiring lock "2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.852200] env[63531]: DEBUG oslo_concurrency.lockutils [None req-795646f8-c711-4669-82d5-9d0c4ed80a2a tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Lock "2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.852301] env[63531]: DEBUG oslo_concurrency.lockutils [None req-795646f8-c711-4669-82d5-9d0c4ed80a2a tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Lock "2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.854874] env[63531]: INFO nova.compute.manager [None req-795646f8-c711-4669-82d5-9d0c4ed80a2a tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Terminating instance [ 973.856835] env[63531]: DEBUG nova.compute.manager [None req-795646f8-c711-4669-82d5-9d0c4ed80a2a tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 973.857049] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-795646f8-c711-4669-82d5-9d0c4ed80a2a tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 973.857928] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba7d60f9-54a6-4e83-98bc-4885df1dd2bd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.867584] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-795646f8-c711-4669-82d5-9d0c4ed80a2a tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 973.867878] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c1cb685e-afe2-4739-b9ac-7dfaf74f9f44 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.875168] env[63531]: DEBUG oslo_vmware.api [None req-795646f8-c711-4669-82d5-9d0c4ed80a2a tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Waiting for the task: (returnval){ [ 973.875168] env[63531]: value = "task-1118298" [ 973.875168] env[63531]: _type = "Task" [ 973.875168] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.883490] env[63531]: DEBUG oslo_vmware.api [None req-795646f8-c711-4669-82d5-9d0c4ed80a2a tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': task-1118298, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.927366] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7f4802a-a604-4325-8e5e-caa6e03d24a9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.947495] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3c84951-0002-433e-9b56-f842112a18da {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.956805] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Updating instance '27c23b7c-a68f-43c9-a125-bd657feb3c5b' progress to 83 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 974.027464] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-666359c3-e81e-471e-99c2-e468372512cc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.034778] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35b7dd0c-8f98-4303-abcd-6b060b2962e2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.065552] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ea12b7c-c681-4ea0-b67d-198c039e887d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.073679] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0204ed07-8475-4798-971b-5e357643e5dd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.093017] env[63531]: DEBUG nova.compute.provider_tree [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 974.216898] env[63531]: DEBUG nova.network.neutron [None req-16b6231f-55ef-4d9c-ab7e-83b12b5da861 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Updating instance_info_cache with network_info: [{"id": "a70530fb-7805-40a0-bea5-a2a666dc679d", "address": "fa:16:3e:78:c3:90", "network": {"id": "73e6d3ed-2016-498a-bda6-fd7dfbff93da", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1501868918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f727ddb992dc4c06954f25f6a1613b4a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c20f5114-0866-45b3-9a7c-62f113ff83fa", "external-id": "nsx-vlan-transportzone-47", "segmentation_id": 47, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa70530fb-78", "ovs_interfaceid": "a70530fb-7805-40a0-bea5-a2a666dc679d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 974.223258] env[63531]: DEBUG oslo_vmware.api [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118297, 'name': PowerOffVM_Task, 'duration_secs': 0.177514} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.223531] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 974.223739] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Updating instance '29dd6de3-2765-408c-acc0-da47e5e0a977' progress to 17 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 974.257413] env[63531]: DEBUG nova.compute.utils [None req-fc9401a8-2566-43ef-b491-055728bc20c4 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 974.384950] env[63531]: DEBUG oslo_vmware.api [None req-795646f8-c711-4669-82d5-9d0c4ed80a2a tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': task-1118298, 'name': PowerOffVM_Task, 'duration_secs': 0.167513} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.385240] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-795646f8-c711-4669-82d5-9d0c4ed80a2a tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 974.385416] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-795646f8-c711-4669-82d5-9d0c4ed80a2a tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 974.385677] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-49d7c799-c48a-4ab5-aa81-609a0595056a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.450877] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-795646f8-c711-4669-82d5-9d0c4ed80a2a tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 974.451109] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-795646f8-c711-4669-82d5-9d0c4ed80a2a tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Deleting contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 974.451240] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-795646f8-c711-4669-82d5-9d0c4ed80a2a tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Deleting the datastore file [datastore2] 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 974.451502] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9f1462fd-cc6f-451e-988a-23d35f83b320 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.458578] env[63531]: DEBUG oslo_vmware.api [None req-795646f8-c711-4669-82d5-9d0c4ed80a2a tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Waiting for the task: (returnval){ [ 974.458578] env[63531]: value = "task-1118300" [ 974.458578] env[63531]: _type = "Task" [ 974.458578] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.466328] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 974.469468] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-748a4604-1060-4b71-83c9-b90d3c12667a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.471076] env[63531]: DEBUG oslo_vmware.api [None req-795646f8-c711-4669-82d5-9d0c4ed80a2a tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': task-1118300, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.475032] env[63531]: DEBUG oslo_vmware.api [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 974.475032] env[63531]: value = "task-1118301" [ 974.475032] env[63531]: _type = "Task" [ 974.475032] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.482017] env[63531]: DEBUG oslo_vmware.api [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118301, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.594896] env[63531]: DEBUG nova.scheduler.client.report [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 974.719737] env[63531]: DEBUG oslo_concurrency.lockutils [None req-16b6231f-55ef-4d9c-ab7e-83b12b5da861 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Releasing lock "refresh_cache-1ca7b851-2a06-4181-8271-58aafcd322d6" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 974.720758] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cde45ce3-99f0-4e83-9f11-3349458f5109 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.729473] env[63531]: DEBUG nova.virt.hardware [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 974.729885] env[63531]: DEBUG nova.virt.hardware [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 974.729885] env[63531]: DEBUG nova.virt.hardware [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 974.730060] env[63531]: DEBUG nova.virt.hardware [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 974.730217] env[63531]: DEBUG nova.virt.hardware [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 974.730385] env[63531]: DEBUG nova.virt.hardware [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 974.730577] env[63531]: DEBUG nova.virt.hardware [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 974.730743] env[63531]: DEBUG nova.virt.hardware [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 974.730919] env[63531]: DEBUG nova.virt.hardware [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 974.731200] env[63531]: DEBUG nova.virt.hardware [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 974.731294] env[63531]: DEBUG nova.virt.hardware [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 974.736163] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-16b6231f-55ef-4d9c-ab7e-83b12b5da861 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Resuming the VM {{(pid=63531) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1184}} [ 974.736414] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-26567848-9696-4ff3-aff7-045c414a66ce {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.746239] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0ffdbc96-7705-40db-a00d-136e88af5f12 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.752928] env[63531]: DEBUG oslo_vmware.api [None req-16b6231f-55ef-4d9c-ab7e-83b12b5da861 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Waiting for the task: (returnval){ [ 974.752928] env[63531]: value = "task-1118303" [ 974.752928] env[63531]: _type = "Task" [ 974.752928] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.757267] env[63531]: DEBUG oslo_vmware.api [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 974.757267] env[63531]: value = "task-1118302" [ 974.757267] env[63531]: _type = "Task" [ 974.757267] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.763905] env[63531]: DEBUG nova.compute.manager [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: 0fb3b687-62cc-4296-b4ed-537aba18c880] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 974.766062] env[63531]: DEBUG oslo_concurrency.lockutils [None req-fc9401a8-2566-43ef-b491-055728bc20c4 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Lock "ccf00b09-29fe-4a76-a8af-97774acd77de" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.012s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.766558] env[63531]: DEBUG oslo_vmware.api [None req-16b6231f-55ef-4d9c-ab7e-83b12b5da861 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1118303, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.772424] env[63531]: DEBUG oslo_vmware.api [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118302, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.794321] env[63531]: DEBUG nova.virt.hardware [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 974.794674] env[63531]: DEBUG nova.virt.hardware [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 974.794875] env[63531]: DEBUG nova.virt.hardware [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 974.795095] env[63531]: DEBUG nova.virt.hardware [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 974.795277] env[63531]: DEBUG nova.virt.hardware [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 974.795439] env[63531]: DEBUG nova.virt.hardware [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 974.795980] env[63531]: DEBUG nova.virt.hardware [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 974.795980] env[63531]: DEBUG nova.virt.hardware [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 974.796138] env[63531]: DEBUG nova.virt.hardware [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 974.796252] env[63531]: DEBUG nova.virt.hardware [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 974.796516] env[63531]: DEBUG nova.virt.hardware [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 974.797449] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-591ef732-1bb3-4c58-8f23-0d8696f87113 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.805847] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85fc9d54-bb95-4033-b5c6-dff05083afa2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.969399] env[63531]: DEBUG oslo_vmware.api [None req-795646f8-c711-4669-82d5-9d0c4ed80a2a tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Task: {'id': task-1118300, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.19164} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.969736] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-795646f8-c711-4669-82d5-9d0c4ed80a2a tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 974.970017] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-795646f8-c711-4669-82d5-9d0c4ed80a2a tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Deleted contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 974.970294] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-795646f8-c711-4669-82d5-9d0c4ed80a2a tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 974.970542] env[63531]: INFO nova.compute.manager [None req-795646f8-c711-4669-82d5-9d0c4ed80a2a tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Took 1.11 seconds to destroy the instance on the hypervisor. [ 974.970845] env[63531]: DEBUG oslo.service.loopingcall [None req-795646f8-c711-4669-82d5-9d0c4ed80a2a tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 974.971178] env[63531]: DEBUG nova.compute.manager [-] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 974.971292] env[63531]: DEBUG nova.network.neutron [-] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 974.984532] env[63531]: DEBUG oslo_vmware.api [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118301, 'name': PowerOnVM_Task, 'duration_secs': 0.407808} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.984829] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 974.985035] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-47231c64-c986-47ba-9b60-40b8945ad3a3 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Updating instance '27c23b7c-a68f-43c9-a125-bd657feb3c5b' progress to 100 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 975.100536] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.368s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.101584] env[63531]: DEBUG nova.compute.manager [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: baf2db4e-2cdb-47f5-9a5c-97a233a459c9] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 975.104228] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.494s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.105977] env[63531]: INFO nova.compute.claims [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3bf08a09-b6e6-4a58-8d35-7061a8e59afd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 975.269183] env[63531]: DEBUG oslo_vmware.api [None req-16b6231f-55ef-4d9c-ab7e-83b12b5da861 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1118303, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.276199] env[63531]: DEBUG oslo_vmware.api [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118302, 'name': ReconfigVM_Task, 'duration_secs': 0.214216} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.278425] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Updating instance '29dd6de3-2765-408c-acc0-da47e5e0a977' progress to 33 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 975.362148] env[63531]: DEBUG nova.compute.manager [req-7e9db9a9-0eb1-4823-a929-211a6fc45956 req-250d31c9-2af4-4408-989d-4f8df9fa2b51 service nova] [instance: 0fb3b687-62cc-4296-b4ed-537aba18c880] Received event network-vif-plugged-ac505e2f-66a8-467d-b109-d86556eedc79 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 975.362309] env[63531]: DEBUG oslo_concurrency.lockutils [req-7e9db9a9-0eb1-4823-a929-211a6fc45956 req-250d31c9-2af4-4408-989d-4f8df9fa2b51 service nova] Acquiring lock "0fb3b687-62cc-4296-b4ed-537aba18c880-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.362520] env[63531]: DEBUG oslo_concurrency.lockutils [req-7e9db9a9-0eb1-4823-a929-211a6fc45956 req-250d31c9-2af4-4408-989d-4f8df9fa2b51 service nova] Lock "0fb3b687-62cc-4296-b4ed-537aba18c880-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.362703] env[63531]: DEBUG oslo_concurrency.lockutils [req-7e9db9a9-0eb1-4823-a929-211a6fc45956 req-250d31c9-2af4-4408-989d-4f8df9fa2b51 service nova] Lock "0fb3b687-62cc-4296-b4ed-537aba18c880-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.362892] env[63531]: DEBUG nova.compute.manager [req-7e9db9a9-0eb1-4823-a929-211a6fc45956 req-250d31c9-2af4-4408-989d-4f8df9fa2b51 service nova] [instance: 0fb3b687-62cc-4296-b4ed-537aba18c880] No waiting events found dispatching network-vif-plugged-ac505e2f-66a8-467d-b109-d86556eedc79 {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 975.364011] env[63531]: WARNING nova.compute.manager [req-7e9db9a9-0eb1-4823-a929-211a6fc45956 req-250d31c9-2af4-4408-989d-4f8df9fa2b51 service nova] [instance: 0fb3b687-62cc-4296-b4ed-537aba18c880] Received unexpected event network-vif-plugged-ac505e2f-66a8-467d-b109-d86556eedc79 for instance with vm_state building and task_state spawning. [ 975.610470] env[63531]: DEBUG nova.compute.manager [req-6b0b9f7c-986e-445d-b558-1ad9a289e223 req-098bc646-3fc6-4c76-a929-730db29b6cca service nova] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Received event network-vif-deleted-2d27a8a5-c288-44fe-ac09-0d3b488386d6 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 975.610470] env[63531]: INFO nova.compute.manager [req-6b0b9f7c-986e-445d-b558-1ad9a289e223 req-098bc646-3fc6-4c76-a929-730db29b6cca service nova] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Neutron deleted interface 2d27a8a5-c288-44fe-ac09-0d3b488386d6; detaching it from the instance and deleting it from the info cache [ 975.610470] env[63531]: DEBUG nova.network.neutron [req-6b0b9f7c-986e-445d-b558-1ad9a289e223 req-098bc646-3fc6-4c76-a929-730db29b6cca service nova] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 975.612660] env[63531]: DEBUG nova.compute.utils [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 975.617037] env[63531]: DEBUG nova.compute.manager [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: baf2db4e-2cdb-47f5-9a5c-97a233a459c9] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 975.617037] env[63531]: DEBUG nova.network.neutron [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: baf2db4e-2cdb-47f5-9a5c-97a233a459c9] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 975.659526] env[63531]: DEBUG nova.policy [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4186052665ce457b99e1fee971be05b4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4ab5ee56531c4e638f8b695b9e5d082f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 975.751734] env[63531]: DEBUG nova.network.neutron [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: 0fb3b687-62cc-4296-b4ed-537aba18c880] Successfully updated port: ac505e2f-66a8-467d-b109-d86556eedc79 {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 975.765799] env[63531]: DEBUG oslo_vmware.api [None req-16b6231f-55ef-4d9c-ab7e-83b12b5da861 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1118303, 'name': PowerOnVM_Task, 'duration_secs': 0.57489} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.767044] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-16b6231f-55ef-4d9c-ab7e-83b12b5da861 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Resumed the VM {{(pid=63531) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1189}} [ 975.768354] env[63531]: DEBUG nova.compute.manager [None req-16b6231f-55ef-4d9c-ab7e-83b12b5da861 tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 975.770424] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38476ad9-929c-4393-a5f4-a8d7bab7a785 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.786234] env[63531]: DEBUG nova.virt.hardware [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 975.787438] env[63531]: DEBUG nova.virt.hardware [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 975.787438] env[63531]: DEBUG nova.virt.hardware [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 975.787438] env[63531]: DEBUG nova.virt.hardware [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 975.787438] env[63531]: DEBUG nova.virt.hardware [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 975.787689] env[63531]: DEBUG nova.virt.hardware [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 975.788194] env[63531]: DEBUG nova.virt.hardware [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 975.788469] env[63531]: DEBUG nova.virt.hardware [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 975.788782] env[63531]: DEBUG nova.virt.hardware [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 975.792021] env[63531]: DEBUG nova.virt.hardware [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 975.792021] env[63531]: DEBUG nova.virt.hardware [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 975.795537] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Reconfiguring VM instance instance-0000002f to detach disk 2000 {{(pid=63531) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 975.796653] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1c6a3b34-78db-4d33-a4ab-d9d70977c328 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.821692] env[63531]: DEBUG oslo_vmware.api [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 975.821692] env[63531]: value = "task-1118305" [ 975.821692] env[63531]: _type = "Task" [ 975.821692] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.829035] env[63531]: DEBUG nova.compute.manager [req-f293c7ee-373f-435a-8be0-acce9276a912 req-e99e8b23-05a4-4e82-beb6-ad23663275d1 service nova] [instance: 0fb3b687-62cc-4296-b4ed-537aba18c880] Received event network-changed-ac505e2f-66a8-467d-b109-d86556eedc79 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 975.829534] env[63531]: DEBUG nova.compute.manager [req-f293c7ee-373f-435a-8be0-acce9276a912 req-e99e8b23-05a4-4e82-beb6-ad23663275d1 service nova] [instance: 0fb3b687-62cc-4296-b4ed-537aba18c880] Refreshing instance network info cache due to event network-changed-ac505e2f-66a8-467d-b109-d86556eedc79. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 975.829841] env[63531]: DEBUG oslo_concurrency.lockutils [req-f293c7ee-373f-435a-8be0-acce9276a912 req-e99e8b23-05a4-4e82-beb6-ad23663275d1 service nova] Acquiring lock "refresh_cache-0fb3b687-62cc-4296-b4ed-537aba18c880" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 975.830230] env[63531]: DEBUG oslo_concurrency.lockutils [req-f293c7ee-373f-435a-8be0-acce9276a912 req-e99e8b23-05a4-4e82-beb6-ad23663275d1 service nova] Acquired lock "refresh_cache-0fb3b687-62cc-4296-b4ed-537aba18c880" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.830427] env[63531]: DEBUG nova.network.neutron [req-f293c7ee-373f-435a-8be0-acce9276a912 req-e99e8b23-05a4-4e82-beb6-ad23663275d1 service nova] [instance: 0fb3b687-62cc-4296-b4ed-537aba18c880] Refreshing network info cache for port ac505e2f-66a8-467d-b109-d86556eedc79 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 975.838795] env[63531]: DEBUG oslo_vmware.api [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118305, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.854016] env[63531]: DEBUG oslo_concurrency.lockutils [None req-fc9401a8-2566-43ef-b491-055728bc20c4 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquiring lock "ccf00b09-29fe-4a76-a8af-97774acd77de" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.854576] env[63531]: DEBUG oslo_concurrency.lockutils [None req-fc9401a8-2566-43ef-b491-055728bc20c4 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Lock "ccf00b09-29fe-4a76-a8af-97774acd77de" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.854903] env[63531]: INFO nova.compute.manager [None req-fc9401a8-2566-43ef-b491-055728bc20c4 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Attaching volume 0bd45432-88d2-467a-aacf-1a9930b4e929 to /dev/sdb [ 975.889642] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2852eb86-860d-4ea4-8bb6-8e3dbdb20c84 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.897653] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1942df77-e111-4c46-a5b5-b30ca77aee1a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.911476] env[63531]: DEBUG nova.virt.block_device [None req-fc9401a8-2566-43ef-b491-055728bc20c4 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Updating existing volume attachment record: e2762f2b-a8c4-44f7-818e-828bb1664f7d {{(pid=63531) _volume_attach /opt/stack/nova/nova/virt/block_device.py:679}} [ 976.069798] env[63531]: DEBUG nova.network.neutron [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: baf2db4e-2cdb-47f5-9a5c-97a233a459c9] Successfully created port: df2cde2b-a781-49ba-9120-bc742e62367d {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 976.089590] env[63531]: DEBUG nova.network.neutron [-] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 976.120102] env[63531]: DEBUG nova.compute.manager [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: baf2db4e-2cdb-47f5-9a5c-97a233a459c9] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 976.124533] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-de1800cc-3734-415f-bccf-fc7ccde2172f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.133108] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-033899a0-a292-4473-b14f-07ffb433c42e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.162164] env[63531]: DEBUG nova.compute.manager [req-6b0b9f7c-986e-445d-b558-1ad9a289e223 req-098bc646-3fc6-4c76-a929-730db29b6cca service nova] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Detach interface failed, port_id=2d27a8a5-c288-44fe-ac09-0d3b488386d6, reason: Instance 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f could not be found. {{(pid=63531) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 976.260682] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Acquiring lock "refresh_cache-0fb3b687-62cc-4296-b4ed-537aba18c880" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 976.331614] env[63531]: DEBUG oslo_vmware.api [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118305, 'name': ReconfigVM_Task, 'duration_secs': 0.163178} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.335341] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Reconfigured VM instance instance-0000002f to detach disk 2000 {{(pid=63531) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 976.337308] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77800da4-0869-4c03-b00c-ebcb9085ff57 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.364692] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Reconfiguring VM instance instance-0000002f to attach disk [datastore2] 29dd6de3-2765-408c-acc0-da47e5e0a977/29dd6de3-2765-408c-acc0-da47e5e0a977.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 976.367484] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d48252ba-1c8c-4082-ad76-8859a81cb104 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.381559] env[63531]: DEBUG nova.network.neutron [req-f293c7ee-373f-435a-8be0-acce9276a912 req-e99e8b23-05a4-4e82-beb6-ad23663275d1 service nova] [instance: 0fb3b687-62cc-4296-b4ed-537aba18c880] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 976.389182] env[63531]: DEBUG oslo_vmware.api [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 976.389182] env[63531]: value = "task-1118309" [ 976.389182] env[63531]: _type = "Task" [ 976.389182] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.399428] env[63531]: DEBUG oslo_vmware.api [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118309, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.480608] env[63531]: DEBUG nova.network.neutron [req-f293c7ee-373f-435a-8be0-acce9276a912 req-e99e8b23-05a4-4e82-beb6-ad23663275d1 service nova] [instance: 0fb3b687-62cc-4296-b4ed-537aba18c880] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 976.541729] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc3f045c-b349-4dcc-83d8-d195d9b1429f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.551303] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dec9559-f552-4624-ba1d-edc69719d047 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.581982] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cc6563a-46a5-4626-b45b-0d8ac373be11 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.588672] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-689472ab-572d-44f5-810e-67f8d09324ce {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.592557] env[63531]: INFO nova.compute.manager [-] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Took 1.62 seconds to deallocate network for instance. [ 976.604549] env[63531]: DEBUG nova.compute.provider_tree [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 976.900603] env[63531]: DEBUG oslo_vmware.api [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118309, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.983220] env[63531]: DEBUG oslo_concurrency.lockutils [req-f293c7ee-373f-435a-8be0-acce9276a912 req-e99e8b23-05a4-4e82-beb6-ad23663275d1 service nova] Releasing lock "refresh_cache-0fb3b687-62cc-4296-b4ed-537aba18c880" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 976.983696] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Acquired lock "refresh_cache-0fb3b687-62cc-4296-b4ed-537aba18c880" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 976.983956] env[63531]: DEBUG nova.network.neutron [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: 0fb3b687-62cc-4296-b4ed-537aba18c880] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 977.013261] env[63531]: DEBUG oslo_concurrency.lockutils [None req-17a8f660-d313-4f51-a9dd-8d78a43ce44c tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "27c23b7c-a68f-43c9-a125-bd657feb3c5b" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.013601] env[63531]: DEBUG oslo_concurrency.lockutils [None req-17a8f660-d313-4f51-a9dd-8d78a43ce44c tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "27c23b7c-a68f-43c9-a125-bd657feb3c5b" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.013850] env[63531]: DEBUG nova.compute.manager [None req-17a8f660-d313-4f51-a9dd-8d78a43ce44c tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Going to confirm migration 3 {{(pid=63531) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 977.107485] env[63531]: DEBUG oslo_concurrency.lockutils [None req-795646f8-c711-4669-82d5-9d0c4ed80a2a tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.111501] env[63531]: DEBUG nova.scheduler.client.report [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 977.133072] env[63531]: DEBUG nova.compute.manager [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: baf2db4e-2cdb-47f5-9a5c-97a233a459c9] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 977.161643] env[63531]: DEBUG nova.virt.hardware [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 977.161998] env[63531]: DEBUG nova.virt.hardware [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 977.162353] env[63531]: DEBUG nova.virt.hardware [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 977.162614] env[63531]: DEBUG nova.virt.hardware [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 977.162754] env[63531]: DEBUG nova.virt.hardware [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 977.162960] env[63531]: DEBUG nova.virt.hardware [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 977.163261] env[63531]: DEBUG nova.virt.hardware [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 977.163440] env[63531]: DEBUG nova.virt.hardware [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 977.163631] env[63531]: DEBUG nova.virt.hardware [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 977.163870] env[63531]: DEBUG nova.virt.hardware [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 977.164090] env[63531]: DEBUG nova.virt.hardware [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 977.165477] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bab4132-c231-4590-a94e-76c16a260392 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.174845] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cda0a4c-6a36-427a-8435-662532ce7326 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.399493] env[63531]: DEBUG oslo_vmware.api [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118309, 'name': ReconfigVM_Task, 'duration_secs': 0.839055} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.399792] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Reconfigured VM instance instance-0000002f to attach disk [datastore2] 29dd6de3-2765-408c-acc0-da47e5e0a977/29dd6de3-2765-408c-acc0-da47e5e0a977.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 977.400107] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Updating instance '29dd6de3-2765-408c-acc0-da47e5e0a977' progress to 50 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 977.525393] env[63531]: DEBUG nova.network.neutron [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: 0fb3b687-62cc-4296-b4ed-537aba18c880] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 977.614790] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-d93e69f3-6cef-48fe-bf5f-c50010578386 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Volume attach. Driver type: vmdk {{(pid=63531) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 977.615053] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-d93e69f3-6cef-48fe-bf5f-c50010578386 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244765', 'volume_id': '332ac07f-9532-470d-aae2-732a5fea0e2a', 'name': 'volume-332ac07f-9532-470d-aae2-732a5fea0e2a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe', 'attached_at': '', 'detached_at': '', 'volume_id': '332ac07f-9532-470d-aae2-732a5fea0e2a', 'serial': '332ac07f-9532-470d-aae2-732a5fea0e2a'} {{(pid=63531) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 977.615872] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.512s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.616375] env[63531]: DEBUG nova.compute.manager [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3bf08a09-b6e6-4a58-8d35-7061a8e59afd] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 977.625023] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef4722dd-a01b-47c2-8053-6cb3625f3811 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.633324] env[63531]: DEBUG oslo_concurrency.lockutils [None req-17a8f660-d313-4f51-a9dd-8d78a43ce44c tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "refresh_cache-27c23b7c-a68f-43c9-a125-bd657feb3c5b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 977.634288] env[63531]: DEBUG oslo_concurrency.lockutils [None req-17a8f660-d313-4f51-a9dd-8d78a43ce44c tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquired lock "refresh_cache-27c23b7c-a68f-43c9-a125-bd657feb3c5b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.634288] env[63531]: DEBUG nova.network.neutron [None req-17a8f660-d313-4f51-a9dd-8d78a43ce44c tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 977.634288] env[63531]: DEBUG nova.objects.instance [None req-17a8f660-d313-4f51-a9dd-8d78a43ce44c tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lazy-loading 'info_cache' on Instance uuid 27c23b7c-a68f-43c9-a125-bd657feb3c5b {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 977.639916] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a384b8d7-53a1-4e7f-90b4-de586fca38e3 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.977s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.640246] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a384b8d7-53a1-4e7f-90b4-de586fca38e3 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.643639] env[63531]: DEBUG oslo_concurrency.lockutils [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.028s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.645823] env[63531]: INFO nova.compute.claims [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 977.653522] env[63531]: DEBUG nova.compute.manager [req-5f5faab5-24b3-4b57-8c81-07faed6f94c7 req-4e9ee851-bad2-4f6d-a10c-1b30086c0946 service nova] [instance: baf2db4e-2cdb-47f5-9a5c-97a233a459c9] Received event network-vif-plugged-df2cde2b-a781-49ba-9120-bc742e62367d {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 977.655382] env[63531]: DEBUG oslo_concurrency.lockutils [req-5f5faab5-24b3-4b57-8c81-07faed6f94c7 req-4e9ee851-bad2-4f6d-a10c-1b30086c0946 service nova] Acquiring lock "baf2db4e-2cdb-47f5-9a5c-97a233a459c9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.655382] env[63531]: DEBUG oslo_concurrency.lockutils [req-5f5faab5-24b3-4b57-8c81-07faed6f94c7 req-4e9ee851-bad2-4f6d-a10c-1b30086c0946 service nova] Lock "baf2db4e-2cdb-47f5-9a5c-97a233a459c9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.655382] env[63531]: DEBUG oslo_concurrency.lockutils [req-5f5faab5-24b3-4b57-8c81-07faed6f94c7 req-4e9ee851-bad2-4f6d-a10c-1b30086c0946 service nova] Lock "baf2db4e-2cdb-47f5-9a5c-97a233a459c9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.655382] env[63531]: DEBUG nova.compute.manager [req-5f5faab5-24b3-4b57-8c81-07faed6f94c7 req-4e9ee851-bad2-4f6d-a10c-1b30086c0946 service nova] [instance: baf2db4e-2cdb-47f5-9a5c-97a233a459c9] No waiting events found dispatching network-vif-plugged-df2cde2b-a781-49ba-9120-bc742e62367d {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 977.655382] env[63531]: WARNING nova.compute.manager [req-5f5faab5-24b3-4b57-8c81-07faed6f94c7 req-4e9ee851-bad2-4f6d-a10c-1b30086c0946 service nova] [instance: baf2db4e-2cdb-47f5-9a5c-97a233a459c9] Received unexpected event network-vif-plugged-df2cde2b-a781-49ba-9120-bc742e62367d for instance with vm_state building and task_state spawning. [ 977.673661] env[63531]: INFO nova.scheduler.client.report [None req-a384b8d7-53a1-4e7f-90b4-de586fca38e3 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Deleted allocations for instance 9a3fad46-a15a-451c-bdab-a3c8cc8add07 [ 977.675260] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8314143-6d86-476c-9927-37f08ba2eccd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.706187] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-d93e69f3-6cef-48fe-bf5f-c50010578386 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Reconfiguring VM instance instance-00000056 to attach disk [datastore1] volume-332ac07f-9532-470d-aae2-732a5fea0e2a/volume-332ac07f-9532-470d-aae2-732a5fea0e2a.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 977.712324] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dd8f72bc-5749-4dc3-a43a-d5caa84b30f8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.728713] env[63531]: DEBUG oslo_vmware.api [None req-d93e69f3-6cef-48fe-bf5f-c50010578386 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 977.728713] env[63531]: value = "task-1118310" [ 977.728713] env[63531]: _type = "Task" [ 977.728713] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.737809] env[63531]: DEBUG oslo_vmware.api [None req-d93e69f3-6cef-48fe-bf5f-c50010578386 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118310, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.757188] env[63531]: DEBUG nova.network.neutron [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: 0fb3b687-62cc-4296-b4ed-537aba18c880] Updating instance_info_cache with network_info: [{"id": "ac505e2f-66a8-467d-b109-d86556eedc79", "address": "fa:16:3e:f3:0b:ee", "network": {"id": "66c17d79-e8ec-4c9e-b6e3-5573d8902c4d", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-285539233-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ab5ee56531c4e638f8b695b9e5d082f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5446413d-c3b0-4cd2-a962-62240db178ac", "external-id": "nsx-vlan-transportzone-528", "segmentation_id": 528, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapac505e2f-66", "ovs_interfaceid": "ac505e2f-66a8-467d-b109-d86556eedc79", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 977.797623] env[63531]: DEBUG nova.network.neutron [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: baf2db4e-2cdb-47f5-9a5c-97a233a459c9] Successfully updated port: df2cde2b-a781-49ba-9120-bc742e62367d {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 977.906917] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fc06320-5e1c-4aa2-a661-77a71b64f1a6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.926901] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55cf4343-f786-4b8e-88a4-727af3f7f440 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.946525] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Updating instance '29dd6de3-2765-408c-acc0-da47e5e0a977' progress to 67 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 978.141570] env[63531]: DEBUG nova.compute.utils [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 978.143046] env[63531]: DEBUG nova.compute.manager [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3bf08a09-b6e6-4a58-8d35-7061a8e59afd] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 978.143237] env[63531]: DEBUG nova.network.neutron [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3bf08a09-b6e6-4a58-8d35-7061a8e59afd] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 978.188639] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a384b8d7-53a1-4e7f-90b4-de586fca38e3 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "9a3fad46-a15a-451c-bdab-a3c8cc8add07" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.736s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.216405] env[63531]: DEBUG nova.policy [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7140e12106504b3e95ff49786361771b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '63900312a30b4942854f552758c039cf', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 978.240262] env[63531]: DEBUG oslo_vmware.api [None req-d93e69f3-6cef-48fe-bf5f-c50010578386 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118310, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.260162] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Releasing lock "refresh_cache-0fb3b687-62cc-4296-b4ed-537aba18c880" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 978.260489] env[63531]: DEBUG nova.compute.manager [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: 0fb3b687-62cc-4296-b4ed-537aba18c880] Instance network_info: |[{"id": "ac505e2f-66a8-467d-b109-d86556eedc79", "address": "fa:16:3e:f3:0b:ee", "network": {"id": "66c17d79-e8ec-4c9e-b6e3-5573d8902c4d", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-285539233-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ab5ee56531c4e638f8b695b9e5d082f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5446413d-c3b0-4cd2-a962-62240db178ac", "external-id": "nsx-vlan-transportzone-528", "segmentation_id": 528, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapac505e2f-66", "ovs_interfaceid": "ac505e2f-66a8-467d-b109-d86556eedc79", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 978.260915] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: 0fb3b687-62cc-4296-b4ed-537aba18c880] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f3:0b:ee', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5446413d-c3b0-4cd2-a962-62240db178ac', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ac505e2f-66a8-467d-b109-d86556eedc79', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 978.268371] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Creating folder: Project (4ab5ee56531c4e638f8b695b9e5d082f). Parent ref: group-v244585. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 978.268903] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-864222d1-015b-4c53-aab6-6881216ece1e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.281458] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Created folder: Project (4ab5ee56531c4e638f8b695b9e5d082f) in parent group-v244585. [ 978.281643] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Creating folder: Instances. Parent ref: group-v244768. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 978.281871] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3c32fc6c-9529-4ce2-b96e-1d1fb407ea93 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.290250] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Created folder: Instances in parent group-v244768. [ 978.290466] env[63531]: DEBUG oslo.service.loopingcall [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 978.290653] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0fb3b687-62cc-4296-b4ed-537aba18c880] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 978.290844] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7d5d416f-dccc-4c2c-8be9-02012f6e0db7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.308620] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Acquiring lock "refresh_cache-baf2db4e-2cdb-47f5-9a5c-97a233a459c9" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.308763] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Acquired lock "refresh_cache-baf2db4e-2cdb-47f5-9a5c-97a233a459c9" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.308936] env[63531]: DEBUG nova.network.neutron [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: baf2db4e-2cdb-47f5-9a5c-97a233a459c9] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 978.316989] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 978.316989] env[63531]: value = "task-1118314" [ 978.316989] env[63531]: _type = "Task" [ 978.316989] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.326123] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118314, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.386269] env[63531]: DEBUG nova.network.neutron [None req-17a8f660-d313-4f51-a9dd-8d78a43ce44c tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Updating instance_info_cache with network_info: [{"id": "a755a88f-c86c-4183-b9fa-a6133060ea5e", "address": "fa:16:3e:26:e6:a5", "network": {"id": "bf99b8cb-9efb-4a74-9db1-0f5f0875f26a", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-998205595-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2e99349303e413b9f7896d449cb7dff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc6e6fe1-c4f5-4389-a49f-0978060eebb4", "external-id": "nsx-vlan-transportzone-829", "segmentation_id": 829, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa755a88f-c8", "ovs_interfaceid": "a755a88f-c86c-4183-b9fa-a6133060ea5e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 978.490105] env[63531]: DEBUG nova.network.neutron [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Port 262abfb7-7bcf-40e6-9ce3-b502de45b3b4 binding to destination host cpu-1 is already ACTIVE {{(pid=63531) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 978.647191] env[63531]: DEBUG nova.compute.manager [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3bf08a09-b6e6-4a58-8d35-7061a8e59afd] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 978.687268] env[63531]: DEBUG nova.network.neutron [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3bf08a09-b6e6-4a58-8d35-7061a8e59afd] Successfully created port: aa2bee40-6eaf-4fa4-a313-0cf4ed60310d {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 978.743091] env[63531]: DEBUG oslo_vmware.api [None req-d93e69f3-6cef-48fe-bf5f-c50010578386 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118310, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.827455] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118314, 'name': CreateVM_Task} progress is 25%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.851114] env[63531]: DEBUG nova.network.neutron [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: baf2db4e-2cdb-47f5-9a5c-97a233a459c9] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 978.893550] env[63531]: DEBUG oslo_concurrency.lockutils [None req-17a8f660-d313-4f51-a9dd-8d78a43ce44c tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Releasing lock "refresh_cache-27c23b7c-a68f-43c9-a125-bd657feb3c5b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 978.893550] env[63531]: DEBUG nova.objects.instance [None req-17a8f660-d313-4f51-a9dd-8d78a43ce44c tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lazy-loading 'migration_context' on Instance uuid 27c23b7c-a68f-43c9-a125-bd657feb3c5b {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 978.984111] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ef4f313-43fe-480f-8ca9-50a431481b14 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.991873] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c634f222-de8f-4a7b-a7c4-587fe33e6dd7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.029039] env[63531]: DEBUG nova.network.neutron [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: baf2db4e-2cdb-47f5-9a5c-97a233a459c9] Updating instance_info_cache with network_info: [{"id": "df2cde2b-a781-49ba-9120-bc742e62367d", "address": "fa:16:3e:cd:5e:1f", "network": {"id": "66c17d79-e8ec-4c9e-b6e3-5573d8902c4d", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-285539233-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ab5ee56531c4e638f8b695b9e5d082f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5446413d-c3b0-4cd2-a962-62240db178ac", "external-id": "nsx-vlan-transportzone-528", "segmentation_id": 528, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdf2cde2b-a7", "ovs_interfaceid": "df2cde2b-a781-49ba-9120-bc742e62367d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 979.030588] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a12da38a-5216-4226-964c-1a34f1ded0af {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.040402] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82ed6f95-f3c3-458a-9082-cfa64af48c2e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.058499] env[63531]: DEBUG nova.compute.provider_tree [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 979.246655] env[63531]: DEBUG oslo_vmware.api [None req-d93e69f3-6cef-48fe-bf5f-c50010578386 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118310, 'name': ReconfigVM_Task, 'duration_secs': 1.342079} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.249577] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-d93e69f3-6cef-48fe-bf5f-c50010578386 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Reconfigured VM instance instance-00000056 to attach disk [datastore1] volume-332ac07f-9532-470d-aae2-732a5fea0e2a/volume-332ac07f-9532-470d-aae2-732a5fea0e2a.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 979.252277] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1c3cecd2-0020-493d-8a6a-2ef1990987dd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.266827] env[63531]: DEBUG oslo_vmware.api [None req-d93e69f3-6cef-48fe-bf5f-c50010578386 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 979.266827] env[63531]: value = "task-1118315" [ 979.266827] env[63531]: _type = "Task" [ 979.266827] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.275273] env[63531]: DEBUG oslo_vmware.api [None req-d93e69f3-6cef-48fe-bf5f-c50010578386 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118315, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.328033] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118314, 'name': CreateVM_Task, 'duration_secs': 0.728124} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.328195] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0fb3b687-62cc-4296-b4ed-537aba18c880] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 979.328924] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 979.329101] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.329430] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 979.329681] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a810db60-4f17-4e21-a371-a8fbea05ad72 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.335087] env[63531]: DEBUG oslo_vmware.api [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Waiting for the task: (returnval){ [ 979.335087] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5225a657-3818-66b5-75d7-0bb890fdffa0" [ 979.335087] env[63531]: _type = "Task" [ 979.335087] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.343803] env[63531]: DEBUG oslo_vmware.api [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5225a657-3818-66b5-75d7-0bb890fdffa0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.396229] env[63531]: DEBUG nova.objects.base [None req-17a8f660-d313-4f51-a9dd-8d78a43ce44c tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Object Instance<27c23b7c-a68f-43c9-a125-bd657feb3c5b> lazy-loaded attributes: info_cache,migration_context {{(pid=63531) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 979.397096] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4158e9e-74c9-4b2f-bb3b-56bc1cc7a746 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.417481] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-86ac8f29-fbd0-40b9-bd60-a4622896997d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.423240] env[63531]: DEBUG oslo_vmware.api [None req-17a8f660-d313-4f51-a9dd-8d78a43ce44c tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 979.423240] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5281bebe-02ef-c72a-09c1-2d0cee45e73c" [ 979.423240] env[63531]: _type = "Task" [ 979.423240] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.433151] env[63531]: DEBUG oslo_vmware.api [None req-17a8f660-d313-4f51-a9dd-8d78a43ce44c tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5281bebe-02ef-c72a-09c1-2d0cee45e73c, 'name': SearchDatastore_Task, 'duration_secs': 0.006599} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.433446] env[63531]: DEBUG oslo_concurrency.lockutils [None req-17a8f660-d313-4f51-a9dd-8d78a43ce44c tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.514699] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "29dd6de3-2765-408c-acc0-da47e5e0a977-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.514898] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "29dd6de3-2765-408c-acc0-da47e5e0a977-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.515092] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "29dd6de3-2765-408c-acc0-da47e5e0a977-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.534871] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Releasing lock "refresh_cache-baf2db4e-2cdb-47f5-9a5c-97a233a459c9" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.535198] env[63531]: DEBUG nova.compute.manager [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: baf2db4e-2cdb-47f5-9a5c-97a233a459c9] Instance network_info: |[{"id": "df2cde2b-a781-49ba-9120-bc742e62367d", "address": "fa:16:3e:cd:5e:1f", "network": {"id": "66c17d79-e8ec-4c9e-b6e3-5573d8902c4d", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-285539233-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ab5ee56531c4e638f8b695b9e5d082f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5446413d-c3b0-4cd2-a962-62240db178ac", "external-id": "nsx-vlan-transportzone-528", "segmentation_id": 528, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdf2cde2b-a7", "ovs_interfaceid": "df2cde2b-a781-49ba-9120-bc742e62367d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 979.535837] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: baf2db4e-2cdb-47f5-9a5c-97a233a459c9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cd:5e:1f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5446413d-c3b0-4cd2-a962-62240db178ac', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'df2cde2b-a781-49ba-9120-bc742e62367d', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 979.543220] env[63531]: DEBUG oslo.service.loopingcall [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 979.543669] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: baf2db4e-2cdb-47f5-9a5c-97a233a459c9] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 979.543909] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4dfcdf15-fc7f-4a7c-8710-7b501342c7e2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.560820] env[63531]: DEBUG nova.scheduler.client.report [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 979.568017] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 979.568017] env[63531]: value = "task-1118316" [ 979.568017] env[63531]: _type = "Task" [ 979.568017] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.573834] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118316, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.660190] env[63531]: DEBUG nova.compute.manager [req-4174d065-dc18-4b89-9481-f22a6dbad6dc req-32270cf5-1452-4a0b-8ea8-b6609420aa4f service nova] [instance: baf2db4e-2cdb-47f5-9a5c-97a233a459c9] Received event network-changed-df2cde2b-a781-49ba-9120-bc742e62367d {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 979.660400] env[63531]: DEBUG nova.compute.manager [req-4174d065-dc18-4b89-9481-f22a6dbad6dc req-32270cf5-1452-4a0b-8ea8-b6609420aa4f service nova] [instance: baf2db4e-2cdb-47f5-9a5c-97a233a459c9] Refreshing instance network info cache due to event network-changed-df2cde2b-a781-49ba-9120-bc742e62367d. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 979.660620] env[63531]: DEBUG oslo_concurrency.lockutils [req-4174d065-dc18-4b89-9481-f22a6dbad6dc req-32270cf5-1452-4a0b-8ea8-b6609420aa4f service nova] Acquiring lock "refresh_cache-baf2db4e-2cdb-47f5-9a5c-97a233a459c9" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 979.660994] env[63531]: DEBUG oslo_concurrency.lockutils [req-4174d065-dc18-4b89-9481-f22a6dbad6dc req-32270cf5-1452-4a0b-8ea8-b6609420aa4f service nova] Acquired lock "refresh_cache-baf2db4e-2cdb-47f5-9a5c-97a233a459c9" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.660994] env[63531]: DEBUG nova.network.neutron [req-4174d065-dc18-4b89-9481-f22a6dbad6dc req-32270cf5-1452-4a0b-8ea8-b6609420aa4f service nova] [instance: baf2db4e-2cdb-47f5-9a5c-97a233a459c9] Refreshing network info cache for port df2cde2b-a781-49ba-9120-bc742e62367d {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 979.663081] env[63531]: DEBUG nova.compute.manager [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3bf08a09-b6e6-4a58-8d35-7061a8e59afd] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 979.688544] env[63531]: DEBUG nova.virt.hardware [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 979.688818] env[63531]: DEBUG nova.virt.hardware [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 979.689031] env[63531]: DEBUG nova.virt.hardware [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 979.689236] env[63531]: DEBUG nova.virt.hardware [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 979.689391] env[63531]: DEBUG nova.virt.hardware [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 979.689544] env[63531]: DEBUG nova.virt.hardware [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 979.689759] env[63531]: DEBUG nova.virt.hardware [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 979.689927] env[63531]: DEBUG nova.virt.hardware [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 979.690112] env[63531]: DEBUG nova.virt.hardware [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 979.690284] env[63531]: DEBUG nova.virt.hardware [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 979.690465] env[63531]: DEBUG nova.virt.hardware [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 979.691327] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ceaa111-e650-42d0-93d4-0e6d5f9d3388 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.699582] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a17c6176-bc09-473b-9e65-e429836619d0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.776626] env[63531]: DEBUG oslo_vmware.api [None req-d93e69f3-6cef-48fe-bf5f-c50010578386 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118315, 'name': ReconfigVM_Task, 'duration_secs': 0.139676} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.776951] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-d93e69f3-6cef-48fe-bf5f-c50010578386 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244765', 'volume_id': '332ac07f-9532-470d-aae2-732a5fea0e2a', 'name': 'volume-332ac07f-9532-470d-aae2-732a5fea0e2a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe', 'attached_at': '', 'detached_at': '', 'volume_id': '332ac07f-9532-470d-aae2-732a5fea0e2a', 'serial': '332ac07f-9532-470d-aae2-732a5fea0e2a'} {{(pid=63531) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 979.846569] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquiring lock "18057005-1ff1-4e3d-bccc-8c89755d6db7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.846909] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "18057005-1ff1-4e3d-bccc-8c89755d6db7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.853929] env[63531]: DEBUG oslo_vmware.api [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5225a657-3818-66b5-75d7-0bb890fdffa0, 'name': SearchDatastore_Task, 'duration_secs': 0.012761} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.854553] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.854877] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: 0fb3b687-62cc-4296-b4ed-537aba18c880] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 979.855214] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 979.855441] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.855706] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 979.856759] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dae2a7c6-de9f-43ec-a93c-8c7c5b318ab9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.865461] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 979.865711] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 979.866993] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a77d1f56-05c5-4856-8c87-95a53e29f33e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.874062] env[63531]: DEBUG oslo_vmware.api [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Waiting for the task: (returnval){ [ 979.874062] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5283fee0-1365-6662-ccc1-75885e9ca3d9" [ 979.874062] env[63531]: _type = "Task" [ 979.874062] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.884900] env[63531]: DEBUG oslo_vmware.api [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5283fee0-1365-6662-ccc1-75885e9ca3d9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.066747] env[63531]: DEBUG oslo_concurrency.lockutils [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.423s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.067311] env[63531]: DEBUG nova.compute.manager [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 980.070452] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9609debe-f9ca-4390-a57b-86b7f7ea872e tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.877s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.070674] env[63531]: DEBUG nova.objects.instance [None req-9609debe-f9ca-4390-a57b-86b7f7ea872e tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lazy-loading 'resources' on Instance uuid f46d8a85-6ab7-436a-bc5e-40d0f80974e5 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 980.080018] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118316, 'name': CreateVM_Task, 'duration_secs': 0.360764} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.080629] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: baf2db4e-2cdb-47f5-9a5c-97a233a459c9] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 980.081277] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 980.081455] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 980.081764] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 980.082315] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8f4ae0ae-7f2f-44c8-bd66-cb8a9189321f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.086858] env[63531]: DEBUG oslo_vmware.api [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Waiting for the task: (returnval){ [ 980.086858] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]522b5288-cdc1-9c1d-3452-26b59a80bdbe" [ 980.086858] env[63531]: _type = "Task" [ 980.086858] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.093832] env[63531]: DEBUG oslo_vmware.api [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]522b5288-cdc1-9c1d-3452-26b59a80bdbe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.355958] env[63531]: DEBUG nova.compute.manager [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 980.391098] env[63531]: DEBUG nova.network.neutron [req-4174d065-dc18-4b89-9481-f22a6dbad6dc req-32270cf5-1452-4a0b-8ea8-b6609420aa4f service nova] [instance: baf2db4e-2cdb-47f5-9a5c-97a233a459c9] Updated VIF entry in instance network info cache for port df2cde2b-a781-49ba-9120-bc742e62367d. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 980.391616] env[63531]: DEBUG nova.network.neutron [req-4174d065-dc18-4b89-9481-f22a6dbad6dc req-32270cf5-1452-4a0b-8ea8-b6609420aa4f service nova] [instance: baf2db4e-2cdb-47f5-9a5c-97a233a459c9] Updating instance_info_cache with network_info: [{"id": "df2cde2b-a781-49ba-9120-bc742e62367d", "address": "fa:16:3e:cd:5e:1f", "network": {"id": "66c17d79-e8ec-4c9e-b6e3-5573d8902c4d", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-285539233-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ab5ee56531c4e638f8b695b9e5d082f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5446413d-c3b0-4cd2-a962-62240db178ac", "external-id": "nsx-vlan-transportzone-528", "segmentation_id": 528, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdf2cde2b-a7", "ovs_interfaceid": "df2cde2b-a781-49ba-9120-bc742e62367d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 980.394301] env[63531]: DEBUG oslo_vmware.api [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5283fee0-1365-6662-ccc1-75885e9ca3d9, 'name': SearchDatastore_Task, 'duration_secs': 0.011855} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.395372] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-574fddb6-3e1c-484f-ba8f-e5d5697f2f59 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.408117] env[63531]: DEBUG oslo_vmware.api [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Waiting for the task: (returnval){ [ 980.408117] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]525cb41a-a9c9-ae00-4710-122137f840f3" [ 980.408117] env[63531]: _type = "Task" [ 980.408117] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.415890] env[63531]: DEBUG oslo_vmware.api [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]525cb41a-a9c9-ae00-4710-122137f840f3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.465118] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc9401a8-2566-43ef-b491-055728bc20c4 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Volume attach. Driver type: vmdk {{(pid=63531) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 980.465118] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc9401a8-2566-43ef-b491-055728bc20c4 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244767', 'volume_id': '0bd45432-88d2-467a-aacf-1a9930b4e929', 'name': 'volume-0bd45432-88d2-467a-aacf-1a9930b4e929', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ccf00b09-29fe-4a76-a8af-97774acd77de', 'attached_at': '', 'detached_at': '', 'volume_id': '0bd45432-88d2-467a-aacf-1a9930b4e929', 'serial': '0bd45432-88d2-467a-aacf-1a9930b4e929'} {{(pid=63531) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 980.465118] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f30d7f82-36a1-4ec8-ada8-ac5f227b3001 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.484192] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-122da59e-5e8f-4102-8cb2-74b66e61710b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.512913] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc9401a8-2566-43ef-b491-055728bc20c4 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Reconfiguring VM instance instance-00000051 to attach disk [datastore2] volume-0bd45432-88d2-467a-aacf-1a9930b4e929/volume-0bd45432-88d2-467a-aacf-1a9930b4e929.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 980.513249] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-98bef1ef-5cb1-4da6-b545-7064244756b2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.533998] env[63531]: DEBUG oslo_vmware.api [None req-fc9401a8-2566-43ef-b491-055728bc20c4 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 980.533998] env[63531]: value = "task-1118317" [ 980.533998] env[63531]: _type = "Task" [ 980.533998] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.542287] env[63531]: DEBUG oslo_vmware.api [None req-fc9401a8-2566-43ef-b491-055728bc20c4 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118317, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.556164] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "refresh_cache-29dd6de3-2765-408c-acc0-da47e5e0a977" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 980.556431] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquired lock "refresh_cache-29dd6de3-2765-408c-acc0-da47e5e0a977" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 980.556553] env[63531]: DEBUG nova.network.neutron [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 980.576053] env[63531]: DEBUG nova.compute.utils [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 980.579903] env[63531]: DEBUG nova.compute.manager [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 980.580106] env[63531]: DEBUG nova.network.neutron [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 980.596771] env[63531]: DEBUG oslo_vmware.api [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]522b5288-cdc1-9c1d-3452-26b59a80bdbe, 'name': SearchDatastore_Task, 'duration_secs': 0.01161} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.597098] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.597340] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: baf2db4e-2cdb-47f5-9a5c-97a233a459c9] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 980.597590] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 980.597742] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 980.597950] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 980.599009] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d2325a0a-a9a1-478b-998e-a7d380b127f4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.606468] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 980.608280] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 980.608280] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8b4c4ed0-21b5-4ca4-8760-8e39331e89e7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.616052] env[63531]: DEBUG oslo_vmware.api [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Waiting for the task: (returnval){ [ 980.616052] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]522bc68d-b777-aeab-54e2-f8936d266ec1" [ 980.616052] env[63531]: _type = "Task" [ 980.616052] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.623585] env[63531]: DEBUG oslo_vmware.api [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]522bc68d-b777-aeab-54e2-f8936d266ec1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.641119] env[63531]: DEBUG nova.policy [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1ea586d8e4f947d2902c130530d42466', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f9056486e3b640d0bcdb9369ed5d67ad', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 980.812759] env[63531]: DEBUG nova.objects.instance [None req-d93e69f3-6cef-48fe-bf5f-c50010578386 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lazy-loading 'flavor' on Instance uuid 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 980.873851] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.899055] env[63531]: DEBUG oslo_concurrency.lockutils [req-4174d065-dc18-4b89-9481-f22a6dbad6dc req-32270cf5-1452-4a0b-8ea8-b6609420aa4f service nova] Releasing lock "refresh_cache-baf2db4e-2cdb-47f5-9a5c-97a233a459c9" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.899055] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fd7a956-6f73-47a1-a630-f8c2e797a7da {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.912966] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce12ad17-9230-4920-913d-31dc18da4c46 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.923994] env[63531]: DEBUG oslo_vmware.api [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]525cb41a-a9c9-ae00-4710-122137f840f3, 'name': SearchDatastore_Task, 'duration_secs': 0.035607} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.951537] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.952085] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 0fb3b687-62cc-4296-b4ed-537aba18c880/0fb3b687-62cc-4296-b4ed-537aba18c880.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 980.952758] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-02707121-dd86-4f86-b4aa-b25a9cadfb0c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.955864] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e757690b-05b7-4969-aaab-e69cc041f407 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.966188] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e47140d-6f10-4dab-afd3-0c297b07418c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.970329] env[63531]: DEBUG oslo_vmware.api [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Waiting for the task: (returnval){ [ 980.970329] env[63531]: value = "task-1118318" [ 980.970329] env[63531]: _type = "Task" [ 980.970329] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.982295] env[63531]: DEBUG nova.compute.provider_tree [None req-9609debe-f9ca-4390-a57b-86b7f7ea872e tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 980.984319] env[63531]: DEBUG nova.network.neutron [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3bf08a09-b6e6-4a58-8d35-7061a8e59afd] Successfully updated port: aa2bee40-6eaf-4fa4-a313-0cf4ed60310d {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 980.991021] env[63531]: DEBUG oslo_vmware.api [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118318, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.010898] env[63531]: DEBUG nova.network.neutron [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Successfully created port: b37a092f-e03c-417f-bac1-3b2ebb5253d1 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 981.044073] env[63531]: DEBUG oslo_vmware.api [None req-fc9401a8-2566-43ef-b491-055728bc20c4 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118317, 'name': ReconfigVM_Task, 'duration_secs': 0.339586} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.044378] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc9401a8-2566-43ef-b491-055728bc20c4 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Reconfigured VM instance instance-00000051 to attach disk [datastore2] volume-0bd45432-88d2-467a-aacf-1a9930b4e929/volume-0bd45432-88d2-467a-aacf-1a9930b4e929.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 981.049267] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1dae6e85-aeb9-47b0-a409-3c9937965ea3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.066084] env[63531]: DEBUG oslo_vmware.api [None req-fc9401a8-2566-43ef-b491-055728bc20c4 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 981.066084] env[63531]: value = "task-1118319" [ 981.066084] env[63531]: _type = "Task" [ 981.066084] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.074300] env[63531]: DEBUG oslo_vmware.api [None req-fc9401a8-2566-43ef-b491-055728bc20c4 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118319, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.083303] env[63531]: DEBUG nova.compute.manager [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 981.126757] env[63531]: DEBUG oslo_vmware.api [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]522bc68d-b777-aeab-54e2-f8936d266ec1, 'name': SearchDatastore_Task, 'duration_secs': 0.008928} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.127554] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b21757c6-5dd3-49ad-9313-19666d4cdf96 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.132791] env[63531]: DEBUG oslo_vmware.api [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Waiting for the task: (returnval){ [ 981.132791] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]525003b2-39c8-9c87-c8f9-a1d2444546c5" [ 981.132791] env[63531]: _type = "Task" [ 981.132791] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.140873] env[63531]: DEBUG oslo_vmware.api [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]525003b2-39c8-9c87-c8f9-a1d2444546c5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.318476] env[63531]: DEBUG oslo_concurrency.lockutils [None req-d93e69f3-6cef-48fe-bf5f-c50010578386 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.311s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.349146] env[63531]: DEBUG nova.network.neutron [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Updating instance_info_cache with network_info: [{"id": "262abfb7-7bcf-40e6-9ce3-b502de45b3b4", "address": "fa:16:3e:0e:00:c9", "network": {"id": "0667c176-cfdf-4e43-b152-e00bdc224448", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-243715985-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.217", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "308eabafea9c4188a58a0f1c22074d2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8f580e6-1d86-41ee-9ebe-c531cb9299c6", "external-id": "nsx-vlan-transportzone-150", "segmentation_id": 150, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap262abfb7-7b", "ovs_interfaceid": "262abfb7-7bcf-40e6-9ce3-b502de45b3b4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 981.482494] env[63531]: DEBUG oslo_vmware.api [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118318, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.487632] env[63531]: DEBUG nova.scheduler.client.report [None req-9609debe-f9ca-4390-a57b-86b7f7ea872e tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 981.491297] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "refresh_cache-3bf08a09-b6e6-4a58-8d35-7061a8e59afd" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 981.491477] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquired lock "refresh_cache-3bf08a09-b6e6-4a58-8d35-7061a8e59afd" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.491657] env[63531]: DEBUG nova.network.neutron [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3bf08a09-b6e6-4a58-8d35-7061a8e59afd] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 981.527110] env[63531]: DEBUG oslo_concurrency.lockutils [None req-fc3f1fb2-bd92-4abe-bcb8-7277fece7b3a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.527433] env[63531]: DEBUG oslo_concurrency.lockutils [None req-fc3f1fb2-bd92-4abe-bcb8-7277fece7b3a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.528305] env[63531]: DEBUG oslo_concurrency.lockutils [None req-fc3f1fb2-bd92-4abe-bcb8-7277fece7b3a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.528527] env[63531]: DEBUG oslo_concurrency.lockutils [None req-fc3f1fb2-bd92-4abe-bcb8-7277fece7b3a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.528831] env[63531]: DEBUG oslo_concurrency.lockutils [None req-fc3f1fb2-bd92-4abe-bcb8-7277fece7b3a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.532915] env[63531]: INFO nova.compute.manager [None req-fc3f1fb2-bd92-4abe-bcb8-7277fece7b3a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Terminating instance [ 981.535489] env[63531]: DEBUG nova.compute.manager [None req-fc3f1fb2-bd92-4abe-bcb8-7277fece7b3a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 981.535729] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc3f1fb2-bd92-4abe-bcb8-7277fece7b3a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 981.536120] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-99028902-9a09-431c-981f-3a7f3588c635 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.544767] env[63531]: DEBUG oslo_vmware.api [None req-fc3f1fb2-bd92-4abe-bcb8-7277fece7b3a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 981.544767] env[63531]: value = "task-1118320" [ 981.544767] env[63531]: _type = "Task" [ 981.544767] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.555298] env[63531]: DEBUG oslo_vmware.api [None req-fc3f1fb2-bd92-4abe-bcb8-7277fece7b3a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118320, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.576883] env[63531]: DEBUG oslo_vmware.api [None req-fc9401a8-2566-43ef-b491-055728bc20c4 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118319, 'name': ReconfigVM_Task, 'duration_secs': 0.140449} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.578026] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc9401a8-2566-43ef-b491-055728bc20c4 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244767', 'volume_id': '0bd45432-88d2-467a-aacf-1a9930b4e929', 'name': 'volume-0bd45432-88d2-467a-aacf-1a9930b4e929', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ccf00b09-29fe-4a76-a8af-97774acd77de', 'attached_at': '', 'detached_at': '', 'volume_id': '0bd45432-88d2-467a-aacf-1a9930b4e929', 'serial': '0bd45432-88d2-467a-aacf-1a9930b4e929'} {{(pid=63531) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 981.642947] env[63531]: DEBUG oslo_vmware.api [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]525003b2-39c8-9c87-c8f9-a1d2444546c5, 'name': SearchDatastore_Task, 'duration_secs': 0.009201} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.643276] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 981.643522] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] baf2db4e-2cdb-47f5-9a5c-97a233a459c9/baf2db4e-2cdb-47f5-9a5c-97a233a459c9.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 981.643794] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-549f0d60-529e-4772-9c16-02282c193c58 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.653097] env[63531]: DEBUG oslo_vmware.api [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Waiting for the task: (returnval){ [ 981.653097] env[63531]: value = "task-1118321" [ 981.653097] env[63531]: _type = "Task" [ 981.653097] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.664501] env[63531]: DEBUG oslo_vmware.api [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118321, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.730871] env[63531]: DEBUG nova.compute.manager [req-b6491d24-c054-4139-8eb2-456c3e367c39 req-886a2c1b-f1e7-4dd5-899b-288561189a0c service nova] [instance: 3bf08a09-b6e6-4a58-8d35-7061a8e59afd] Received event network-vif-plugged-aa2bee40-6eaf-4fa4-a313-0cf4ed60310d {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 981.731247] env[63531]: DEBUG oslo_concurrency.lockutils [req-b6491d24-c054-4139-8eb2-456c3e367c39 req-886a2c1b-f1e7-4dd5-899b-288561189a0c service nova] Acquiring lock "3bf08a09-b6e6-4a58-8d35-7061a8e59afd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.731445] env[63531]: DEBUG oslo_concurrency.lockutils [req-b6491d24-c054-4139-8eb2-456c3e367c39 req-886a2c1b-f1e7-4dd5-899b-288561189a0c service nova] Lock "3bf08a09-b6e6-4a58-8d35-7061a8e59afd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.731640] env[63531]: DEBUG oslo_concurrency.lockutils [req-b6491d24-c054-4139-8eb2-456c3e367c39 req-886a2c1b-f1e7-4dd5-899b-288561189a0c service nova] Lock "3bf08a09-b6e6-4a58-8d35-7061a8e59afd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.731812] env[63531]: DEBUG nova.compute.manager [req-b6491d24-c054-4139-8eb2-456c3e367c39 req-886a2c1b-f1e7-4dd5-899b-288561189a0c service nova] [instance: 3bf08a09-b6e6-4a58-8d35-7061a8e59afd] No waiting events found dispatching network-vif-plugged-aa2bee40-6eaf-4fa4-a313-0cf4ed60310d {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 981.731991] env[63531]: WARNING nova.compute.manager [req-b6491d24-c054-4139-8eb2-456c3e367c39 req-886a2c1b-f1e7-4dd5-899b-288561189a0c service nova] [instance: 3bf08a09-b6e6-4a58-8d35-7061a8e59afd] Received unexpected event network-vif-plugged-aa2bee40-6eaf-4fa4-a313-0cf4ed60310d for instance with vm_state building and task_state spawning. [ 981.732163] env[63531]: DEBUG nova.compute.manager [req-b6491d24-c054-4139-8eb2-456c3e367c39 req-886a2c1b-f1e7-4dd5-899b-288561189a0c service nova] [instance: 3bf08a09-b6e6-4a58-8d35-7061a8e59afd] Received event network-changed-aa2bee40-6eaf-4fa4-a313-0cf4ed60310d {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 981.732437] env[63531]: DEBUG nova.compute.manager [req-b6491d24-c054-4139-8eb2-456c3e367c39 req-886a2c1b-f1e7-4dd5-899b-288561189a0c service nova] [instance: 3bf08a09-b6e6-4a58-8d35-7061a8e59afd] Refreshing instance network info cache due to event network-changed-aa2bee40-6eaf-4fa4-a313-0cf4ed60310d. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 981.732640] env[63531]: DEBUG oslo_concurrency.lockutils [req-b6491d24-c054-4139-8eb2-456c3e367c39 req-886a2c1b-f1e7-4dd5-899b-288561189a0c service nova] Acquiring lock "refresh_cache-3bf08a09-b6e6-4a58-8d35-7061a8e59afd" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 981.852396] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Releasing lock "refresh_cache-29dd6de3-2765-408c-acc0-da47e5e0a977" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 981.983444] env[63531]: DEBUG oslo_vmware.api [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118318, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.808318} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.983703] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 0fb3b687-62cc-4296-b4ed-537aba18c880/0fb3b687-62cc-4296-b4ed-537aba18c880.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 981.984093] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: 0fb3b687-62cc-4296-b4ed-537aba18c880] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 981.984217] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-91e68b0e-d184-431d-ab1f-054e98db8ca5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.992256] env[63531]: DEBUG oslo_vmware.api [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Waiting for the task: (returnval){ [ 981.992256] env[63531]: value = "task-1118322" [ 981.992256] env[63531]: _type = "Task" [ 981.992256] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.996841] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9609debe-f9ca-4390-a57b-86b7f7ea872e tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.926s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.001128] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.771s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.002856] env[63531]: INFO nova.compute.claims [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 982.011313] env[63531]: DEBUG oslo_vmware.api [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118322, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.021376] env[63531]: INFO nova.scheduler.client.report [None req-9609debe-f9ca-4390-a57b-86b7f7ea872e tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Deleted allocations for instance f46d8a85-6ab7-436a-bc5e-40d0f80974e5 [ 982.058982] env[63531]: DEBUG oslo_vmware.api [None req-fc3f1fb2-bd92-4abe-bcb8-7277fece7b3a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118320, 'name': PowerOffVM_Task, 'duration_secs': 0.213819} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.061864] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc3f1fb2-bd92-4abe-bcb8-7277fece7b3a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 982.062244] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc3f1fb2-bd92-4abe-bcb8-7277fece7b3a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Volume detach. Driver type: vmdk {{(pid=63531) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 982.062499] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc3f1fb2-bd92-4abe-bcb8-7277fece7b3a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244765', 'volume_id': '332ac07f-9532-470d-aae2-732a5fea0e2a', 'name': 'volume-332ac07f-9532-470d-aae2-732a5fea0e2a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe', 'attached_at': '', 'detached_at': '', 'volume_id': '332ac07f-9532-470d-aae2-732a5fea0e2a', 'serial': '332ac07f-9532-470d-aae2-732a5fea0e2a'} {{(pid=63531) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 982.064070] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59d14b77-237a-4399-9925-a6861cbdb1d4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.068186] env[63531]: DEBUG nova.network.neutron [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3bf08a09-b6e6-4a58-8d35-7061a8e59afd] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 982.096134] env[63531]: DEBUG nova.compute.manager [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 982.102250] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-825f4a7f-c1d9-427e-b41b-009c1f02f0ad {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.112369] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3273a15a-96ac-4a50-9758-7f1f8d9dce7d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.137266] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35e3c819-fbf4-4cf9-af46-8fb931fd250f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.157135] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc3f1fb2-bd92-4abe-bcb8-7277fece7b3a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] The volume has not been displaced from its original location: [datastore1] volume-332ac07f-9532-470d-aae2-732a5fea0e2a/volume-332ac07f-9532-470d-aae2-732a5fea0e2a.vmdk. No consolidation needed. {{(pid=63531) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 982.163161] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc3f1fb2-bd92-4abe-bcb8-7277fece7b3a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Reconfiguring VM instance instance-00000056 to detach disk 2001 {{(pid=63531) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 982.168382] env[63531]: DEBUG nova.virt.hardware [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 982.168684] env[63531]: DEBUG nova.virt.hardware [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 982.168920] env[63531]: DEBUG nova.virt.hardware [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 982.169219] env[63531]: DEBUG nova.virt.hardware [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 982.169367] env[63531]: DEBUG nova.virt.hardware [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 982.169527] env[63531]: DEBUG nova.virt.hardware [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 982.171115] env[63531]: DEBUG nova.virt.hardware [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 982.171115] env[63531]: DEBUG nova.virt.hardware [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 982.171115] env[63531]: DEBUG nova.virt.hardware [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 982.171115] env[63531]: DEBUG nova.virt.hardware [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 982.171115] env[63531]: DEBUG nova.virt.hardware [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 982.174567] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b9b64bd0-fc2b-41cc-91c3-311f3c842071 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.188602] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7384ca31-edf1-460b-ae5c-49b064cf6ba9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.202073] env[63531]: DEBUG oslo_vmware.api [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118321, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.202073] env[63531]: DEBUG oslo_vmware.api [None req-fc3f1fb2-bd92-4abe-bcb8-7277fece7b3a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 982.202073] env[63531]: value = "task-1118323" [ 982.202073] env[63531]: _type = "Task" [ 982.202073] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.203176] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e92a650-8571-45e6-b8f7-245bb831a7f1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.210367] env[63531]: DEBUG oslo_concurrency.lockutils [None req-2df5001e-3daa-4d21-82f3-a5fde08b46fb tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Acquiring lock "1ca7b851-2a06-4181-8271-58aafcd322d6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.210572] env[63531]: DEBUG oslo_concurrency.lockutils [None req-2df5001e-3daa-4d21-82f3-a5fde08b46fb tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Lock "1ca7b851-2a06-4181-8271-58aafcd322d6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.210781] env[63531]: DEBUG oslo_concurrency.lockutils [None req-2df5001e-3daa-4d21-82f3-a5fde08b46fb tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Acquiring lock "1ca7b851-2a06-4181-8271-58aafcd322d6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.210965] env[63531]: DEBUG oslo_concurrency.lockutils [None req-2df5001e-3daa-4d21-82f3-a5fde08b46fb tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Lock "1ca7b851-2a06-4181-8271-58aafcd322d6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.211146] env[63531]: DEBUG oslo_concurrency.lockutils [None req-2df5001e-3daa-4d21-82f3-a5fde08b46fb tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Lock "1ca7b851-2a06-4181-8271-58aafcd322d6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.213375] env[63531]: INFO nova.compute.manager [None req-2df5001e-3daa-4d21-82f3-a5fde08b46fb tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Terminating instance [ 982.215543] env[63531]: DEBUG nova.compute.manager [None req-2df5001e-3daa-4d21-82f3-a5fde08b46fb tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 982.215740] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-2df5001e-3daa-4d21-82f3-a5fde08b46fb tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 982.216938] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51b6cdeb-3f5e-4270-b6bb-e37b914bd6cd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.233652] env[63531]: DEBUG oslo_vmware.api [None req-fc3f1fb2-bd92-4abe-bcb8-7277fece7b3a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118323, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.238471] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-2df5001e-3daa-4d21-82f3-a5fde08b46fb tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 982.241410] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-af17b731-865d-48dd-b660-5679d20d2100 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.249150] env[63531]: DEBUG oslo_vmware.api [None req-2df5001e-3daa-4d21-82f3-a5fde08b46fb tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Waiting for the task: (returnval){ [ 982.249150] env[63531]: value = "task-1118324" [ 982.249150] env[63531]: _type = "Task" [ 982.249150] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.256885] env[63531]: DEBUG oslo_vmware.api [None req-2df5001e-3daa-4d21-82f3-a5fde08b46fb tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1118324, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.378942] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72673295-6106-4149-9a35-4829550c0289 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.401838] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d1ff0f2-7b6e-4553-95b3-194018448458 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.405920] env[63531]: DEBUG nova.network.neutron [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3bf08a09-b6e6-4a58-8d35-7061a8e59afd] Updating instance_info_cache with network_info: [{"id": "aa2bee40-6eaf-4fa4-a313-0cf4ed60310d", "address": "fa:16:3e:35:23:9a", "network": {"id": "80641cc0-680b-40ff-8d1b-f5a5f736a88f", "bridge": "br-int", "label": "tempest-ServersTestJSON-666288463-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "63900312a30b4942854f552758c039cf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "26472e27-9835-4f87-ab7f-ca24dfee4e83", "external-id": "nsx-vlan-transportzone-335", "segmentation_id": 335, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaa2bee40-6e", "ovs_interfaceid": "aa2bee40-6eaf-4fa4-a313-0cf4ed60310d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 982.410137] env[63531]: DEBUG nova.compute.manager [req-0b4ee4c4-9341-4004-8d1c-1c86cf66f0b0 req-e5e3f330-3591-47c6-99d8-08ba5c848c02 service nova] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Received event network-vif-plugged-b37a092f-e03c-417f-bac1-3b2ebb5253d1 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 982.410529] env[63531]: DEBUG oslo_concurrency.lockutils [req-0b4ee4c4-9341-4004-8d1c-1c86cf66f0b0 req-e5e3f330-3591-47c6-99d8-08ba5c848c02 service nova] Acquiring lock "4757625a-8398-4b2d-b6fe-40eea8913068-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.410847] env[63531]: DEBUG oslo_concurrency.lockutils [req-0b4ee4c4-9341-4004-8d1c-1c86cf66f0b0 req-e5e3f330-3591-47c6-99d8-08ba5c848c02 service nova] Lock "4757625a-8398-4b2d-b6fe-40eea8913068-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.411258] env[63531]: DEBUG oslo_concurrency.lockutils [req-0b4ee4c4-9341-4004-8d1c-1c86cf66f0b0 req-e5e3f330-3591-47c6-99d8-08ba5c848c02 service nova] Lock "4757625a-8398-4b2d-b6fe-40eea8913068-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.411509] env[63531]: DEBUG nova.compute.manager [req-0b4ee4c4-9341-4004-8d1c-1c86cf66f0b0 req-e5e3f330-3591-47c6-99d8-08ba5c848c02 service nova] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] No waiting events found dispatching network-vif-plugged-b37a092f-e03c-417f-bac1-3b2ebb5253d1 {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 982.411727] env[63531]: WARNING nova.compute.manager [req-0b4ee4c4-9341-4004-8d1c-1c86cf66f0b0 req-e5e3f330-3591-47c6-99d8-08ba5c848c02 service nova] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Received unexpected event network-vif-plugged-b37a092f-e03c-417f-bac1-3b2ebb5253d1 for instance with vm_state building and task_state spawning. [ 982.415166] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Updating instance '29dd6de3-2765-408c-acc0-da47e5e0a977' progress to 83 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 982.502169] env[63531]: DEBUG oslo_vmware.api [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118322, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079924} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.502485] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: 0fb3b687-62cc-4296-b4ed-537aba18c880] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 982.503559] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eded93a6-e2bd-4089-b292-d3743ddeda9d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.525603] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: 0fb3b687-62cc-4296-b4ed-537aba18c880] Reconfiguring VM instance instance-00000057 to attach disk [datastore1] 0fb3b687-62cc-4296-b4ed-537aba18c880/0fb3b687-62cc-4296-b4ed-537aba18c880.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 982.530684] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-45b4eebe-d8f7-4fc0-a3d1-fdc9870bb692 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.544806] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9609debe-f9ca-4390-a57b-86b7f7ea872e tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lock "f46d8a85-6ab7-436a-bc5e-40d0f80974e5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.298s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.551425] env[63531]: DEBUG oslo_vmware.api [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Waiting for the task: (returnval){ [ 982.551425] env[63531]: value = "task-1118325" [ 982.551425] env[63531]: _type = "Task" [ 982.551425] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.561855] env[63531]: DEBUG oslo_vmware.api [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118325, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.644736] env[63531]: DEBUG nova.objects.instance [None req-fc9401a8-2566-43ef-b491-055728bc20c4 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Lazy-loading 'flavor' on Instance uuid ccf00b09-29fe-4a76-a8af-97774acd77de {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 982.668951] env[63531]: DEBUG oslo_vmware.api [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118321, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.670646} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.671288] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] baf2db4e-2cdb-47f5-9a5c-97a233a459c9/baf2db4e-2cdb-47f5-9a5c-97a233a459c9.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 982.671527] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: baf2db4e-2cdb-47f5-9a5c-97a233a459c9] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 982.671949] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7070365f-eb9d-4c09-bc2a-4e403bc925fd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.681466] env[63531]: DEBUG oslo_vmware.api [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Waiting for the task: (returnval){ [ 982.681466] env[63531]: value = "task-1118326" [ 982.681466] env[63531]: _type = "Task" [ 982.681466] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.692098] env[63531]: DEBUG oslo_vmware.api [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118326, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.710730] env[63531]: DEBUG oslo_vmware.api [None req-fc3f1fb2-bd92-4abe-bcb8-7277fece7b3a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118323, 'name': ReconfigVM_Task, 'duration_secs': 0.25084} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.713271] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc3f1fb2-bd92-4abe-bcb8-7277fece7b3a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Reconfigured VM instance instance-00000056 to detach disk 2001 {{(pid=63531) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 982.718064] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2a9bba2e-0a37-4be2-b23c-315fc1831489 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.734828] env[63531]: DEBUG oslo_vmware.api [None req-fc3f1fb2-bd92-4abe-bcb8-7277fece7b3a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 982.734828] env[63531]: value = "task-1118327" [ 982.734828] env[63531]: _type = "Task" [ 982.734828] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.744481] env[63531]: DEBUG oslo_vmware.api [None req-fc3f1fb2-bd92-4abe-bcb8-7277fece7b3a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118327, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.758788] env[63531]: DEBUG oslo_vmware.api [None req-2df5001e-3daa-4d21-82f3-a5fde08b46fb tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1118324, 'name': PowerOffVM_Task, 'duration_secs': 0.210454} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.759119] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-2df5001e-3daa-4d21-82f3-a5fde08b46fb tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 982.759297] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-2df5001e-3daa-4d21-82f3-a5fde08b46fb tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 982.759543] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e22f87ed-29ab-4e9a-b230-d48471a954aa {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.798703] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4e6624e-c47f-4540-a28c-5355ea9e4fbf {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.806364] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93234a07-945f-465d-b8b5-385e0dc55c95 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.836061] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5096493a-8beb-4d56-9398-21e5978fba02 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.844283] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c143177-a7a1-4ca6-829c-74b3d69246e3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.857967] env[63531]: DEBUG nova.compute.provider_tree [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 982.909388] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Releasing lock "refresh_cache-3bf08a09-b6e6-4a58-8d35-7061a8e59afd" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 982.909743] env[63531]: DEBUG nova.compute.manager [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3bf08a09-b6e6-4a58-8d35-7061a8e59afd] Instance network_info: |[{"id": "aa2bee40-6eaf-4fa4-a313-0cf4ed60310d", "address": "fa:16:3e:35:23:9a", "network": {"id": "80641cc0-680b-40ff-8d1b-f5a5f736a88f", "bridge": "br-int", "label": "tempest-ServersTestJSON-666288463-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "63900312a30b4942854f552758c039cf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "26472e27-9835-4f87-ab7f-ca24dfee4e83", "external-id": "nsx-vlan-transportzone-335", "segmentation_id": 335, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaa2bee40-6e", "ovs_interfaceid": "aa2bee40-6eaf-4fa4-a313-0cf4ed60310d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 982.910152] env[63531]: DEBUG oslo_concurrency.lockutils [req-b6491d24-c054-4139-8eb2-456c3e367c39 req-886a2c1b-f1e7-4dd5-899b-288561189a0c service nova] Acquired lock "refresh_cache-3bf08a09-b6e6-4a58-8d35-7061a8e59afd" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.910403] env[63531]: DEBUG nova.network.neutron [req-b6491d24-c054-4139-8eb2-456c3e367c39 req-886a2c1b-f1e7-4dd5-899b-288561189a0c service nova] [instance: 3bf08a09-b6e6-4a58-8d35-7061a8e59afd] Refreshing network info cache for port aa2bee40-6eaf-4fa4-a313-0cf4ed60310d {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 982.911658] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3bf08a09-b6e6-4a58-8d35-7061a8e59afd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:35:23:9a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '26472e27-9835-4f87-ab7f-ca24dfee4e83', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'aa2bee40-6eaf-4fa4-a313-0cf4ed60310d', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 982.919175] env[63531]: DEBUG oslo.service.loopingcall [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 982.922408] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3bf08a09-b6e6-4a58-8d35-7061a8e59afd] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 982.926592] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 982.927523] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5a555d5d-a018-4755-9665-89ff87bcd826 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.943044] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-03a39ca8-b80c-47d8-9689-0bb80f8d2153 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.951633] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 982.951633] env[63531]: value = "task-1118329" [ 982.951633] env[63531]: _type = "Task" [ 982.951633] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.953182] env[63531]: DEBUG oslo_vmware.api [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 982.953182] env[63531]: value = "task-1118330" [ 982.953182] env[63531]: _type = "Task" [ 982.953182] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.964154] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118329, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.967660] env[63531]: DEBUG oslo_vmware.api [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118330, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.977412] env[63531]: DEBUG nova.network.neutron [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Successfully updated port: b37a092f-e03c-417f-bac1-3b2ebb5253d1 {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 983.061532] env[63531]: DEBUG oslo_vmware.api [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118325, 'name': ReconfigVM_Task, 'duration_secs': 0.405954} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.061868] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: 0fb3b687-62cc-4296-b4ed-537aba18c880] Reconfigured VM instance instance-00000057 to attach disk [datastore1] 0fb3b687-62cc-4296-b4ed-537aba18c880/0fb3b687-62cc-4296-b4ed-537aba18c880.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 983.062584] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6e8af360-4be3-4d84-9bc3-a4eeed66566b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.069157] env[63531]: DEBUG oslo_vmware.api [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Waiting for the task: (returnval){ [ 983.069157] env[63531]: value = "task-1118331" [ 983.069157] env[63531]: _type = "Task" [ 983.069157] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.077759] env[63531]: DEBUG oslo_vmware.api [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118331, 'name': Rename_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.116762] env[63531]: DEBUG nova.compute.manager [req-4adebe96-9d08-4dfb-aa3c-33e1f0f4460c req-01cecca0-b420-4eb6-a9dd-80366ab5ade2 service nova] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Received event network-changed-b37a092f-e03c-417f-bac1-3b2ebb5253d1 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 983.117050] env[63531]: DEBUG nova.compute.manager [req-4adebe96-9d08-4dfb-aa3c-33e1f0f4460c req-01cecca0-b420-4eb6-a9dd-80366ab5ade2 service nova] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Refreshing instance network info cache due to event network-changed-b37a092f-e03c-417f-bac1-3b2ebb5253d1. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 983.117319] env[63531]: DEBUG oslo_concurrency.lockutils [req-4adebe96-9d08-4dfb-aa3c-33e1f0f4460c req-01cecca0-b420-4eb6-a9dd-80366ab5ade2 service nova] Acquiring lock "refresh_cache-4757625a-8398-4b2d-b6fe-40eea8913068" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 983.117551] env[63531]: DEBUG oslo_concurrency.lockutils [req-4adebe96-9d08-4dfb-aa3c-33e1f0f4460c req-01cecca0-b420-4eb6-a9dd-80366ab5ade2 service nova] Acquired lock "refresh_cache-4757625a-8398-4b2d-b6fe-40eea8913068" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.117780] env[63531]: DEBUG nova.network.neutron [req-4adebe96-9d08-4dfb-aa3c-33e1f0f4460c req-01cecca0-b420-4eb6-a9dd-80366ab5ade2 service nova] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Refreshing network info cache for port b37a092f-e03c-417f-bac1-3b2ebb5253d1 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 983.150394] env[63531]: DEBUG oslo_concurrency.lockutils [None req-fc9401a8-2566-43ef-b491-055728bc20c4 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Lock "ccf00b09-29fe-4a76-a8af-97774acd77de" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.296s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.191591] env[63531]: DEBUG oslo_vmware.api [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118326, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063758} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.191875] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: baf2db4e-2cdb-47f5-9a5c-97a233a459c9] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 983.192682] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5cf72ec-8dcf-4458-a9ab-0db9685a2e01 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.218425] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: baf2db4e-2cdb-47f5-9a5c-97a233a459c9] Reconfiguring VM instance instance-00000058 to attach disk [datastore2] baf2db4e-2cdb-47f5-9a5c-97a233a459c9/baf2db4e-2cdb-47f5-9a5c-97a233a459c9.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 983.221137] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-27bcd5e0-c538-4942-805b-507544886b85 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.244699] env[63531]: DEBUG oslo_vmware.api [None req-fc3f1fb2-bd92-4abe-bcb8-7277fece7b3a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118327, 'name': ReconfigVM_Task, 'duration_secs': 0.323272} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.246079] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc3f1fb2-bd92-4abe-bcb8-7277fece7b3a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244765', 'volume_id': '332ac07f-9532-470d-aae2-732a5fea0e2a', 'name': 'volume-332ac07f-9532-470d-aae2-732a5fea0e2a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe', 'attached_at': '', 'detached_at': '', 'volume_id': '332ac07f-9532-470d-aae2-732a5fea0e2a', 'serial': '332ac07f-9532-470d-aae2-732a5fea0e2a'} {{(pid=63531) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 983.246405] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-fc3f1fb2-bd92-4abe-bcb8-7277fece7b3a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 983.246728] env[63531]: DEBUG oslo_vmware.api [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Waiting for the task: (returnval){ [ 983.246728] env[63531]: value = "task-1118332" [ 983.246728] env[63531]: _type = "Task" [ 983.246728] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.247440] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-140ffe86-3e3b-43b3-bfa4-b87b08ccf60c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.260363] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-fc3f1fb2-bd92-4abe-bcb8-7277fece7b3a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 983.260713] env[63531]: DEBUG oslo_vmware.api [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118332, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.260953] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4f25f210-3c56-4214-ade4-baa3777ea739 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.360937] env[63531]: DEBUG nova.scheduler.client.report [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 983.397555] env[63531]: DEBUG nova.network.neutron [req-b6491d24-c054-4139-8eb2-456c3e367c39 req-886a2c1b-f1e7-4dd5-899b-288561189a0c service nova] [instance: 3bf08a09-b6e6-4a58-8d35-7061a8e59afd] Updated VIF entry in instance network info cache for port aa2bee40-6eaf-4fa4-a313-0cf4ed60310d. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 983.397965] env[63531]: DEBUG nova.network.neutron [req-b6491d24-c054-4139-8eb2-456c3e367c39 req-886a2c1b-f1e7-4dd5-899b-288561189a0c service nova] [instance: 3bf08a09-b6e6-4a58-8d35-7061a8e59afd] Updating instance_info_cache with network_info: [{"id": "aa2bee40-6eaf-4fa4-a313-0cf4ed60310d", "address": "fa:16:3e:35:23:9a", "network": {"id": "80641cc0-680b-40ff-8d1b-f5a5f736a88f", "bridge": "br-int", "label": "tempest-ServersTestJSON-666288463-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "63900312a30b4942854f552758c039cf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "26472e27-9835-4f87-ab7f-ca24dfee4e83", "external-id": "nsx-vlan-transportzone-335", "segmentation_id": 335, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaa2bee40-6e", "ovs_interfaceid": "aa2bee40-6eaf-4fa4-a313-0cf4ed60310d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.466361] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118329, 'name': CreateVM_Task, 'duration_secs': 0.402331} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.469711] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3bf08a09-b6e6-4a58-8d35-7061a8e59afd] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 983.469920] env[63531]: DEBUG oslo_vmware.api [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118330, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.470646] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 983.470812] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.471154] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 983.471413] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-11352482-301d-4ad3-bef2-892c0c32725e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.476324] env[63531]: DEBUG oslo_vmware.api [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 983.476324] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]524e6de2-9d8a-1a8a-ed22-0d2472dad07f" [ 983.476324] env[63531]: _type = "Task" [ 983.476324] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.481323] env[63531]: DEBUG oslo_concurrency.lockutils [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquiring lock "refresh_cache-4757625a-8398-4b2d-b6fe-40eea8913068" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 983.486315] env[63531]: DEBUG oslo_vmware.api [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]524e6de2-9d8a-1a8a-ed22-0d2472dad07f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.504484] env[63531]: INFO nova.compute.manager [None req-ac717b99-ffee-40ff-8c50-9c164bf5d8ae tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Rescuing [ 983.504768] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ac717b99-ffee-40ff-8c50-9c164bf5d8ae tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquiring lock "refresh_cache-ccf00b09-29fe-4a76-a8af-97774acd77de" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 983.504923] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ac717b99-ffee-40ff-8c50-9c164bf5d8ae tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquired lock "refresh_cache-ccf00b09-29fe-4a76-a8af-97774acd77de" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.505116] env[63531]: DEBUG nova.network.neutron [None req-ac717b99-ffee-40ff-8c50-9c164bf5d8ae tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 983.579977] env[63531]: DEBUG oslo_vmware.api [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118331, 'name': Rename_Task, 'duration_secs': 0.16833} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.579977] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: 0fb3b687-62cc-4296-b4ed-537aba18c880] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 983.579977] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d3c2251f-c653-41d1-903d-e7d52881af4a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.586882] env[63531]: DEBUG oslo_vmware.api [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Waiting for the task: (returnval){ [ 983.586882] env[63531]: value = "task-1118334" [ 983.586882] env[63531]: _type = "Task" [ 983.586882] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.595747] env[63531]: DEBUG oslo_vmware.api [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118334, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.622134] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-fc3f1fb2-bd92-4abe-bcb8-7277fece7b3a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 983.622417] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-fc3f1fb2-bd92-4abe-bcb8-7277fece7b3a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 983.622529] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc3f1fb2-bd92-4abe-bcb8-7277fece7b3a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Deleting the datastore file [datastore1] 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 983.623434] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-478e4413-c471-4486-a8a0-bff33282b4b7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.625212] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-2df5001e-3daa-4d21-82f3-a5fde08b46fb tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 983.625410] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-2df5001e-3daa-4d21-82f3-a5fde08b46fb tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 983.625587] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-2df5001e-3daa-4d21-82f3-a5fde08b46fb tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Deleting the datastore file [datastore1] 1ca7b851-2a06-4181-8271-58aafcd322d6 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 983.626159] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-59e3ec70-f8ac-4aad-bcc6-c8e560f8387b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.632081] env[63531]: DEBUG oslo_vmware.api [None req-fc3f1fb2-bd92-4abe-bcb8-7277fece7b3a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 983.632081] env[63531]: value = "task-1118335" [ 983.632081] env[63531]: _type = "Task" [ 983.632081] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.633396] env[63531]: DEBUG oslo_vmware.api [None req-2df5001e-3daa-4d21-82f3-a5fde08b46fb tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Waiting for the task: (returnval){ [ 983.633396] env[63531]: value = "task-1118336" [ 983.633396] env[63531]: _type = "Task" [ 983.633396] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.648577] env[63531]: DEBUG oslo_vmware.api [None req-fc3f1fb2-bd92-4abe-bcb8-7277fece7b3a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118335, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.652606] env[63531]: DEBUG oslo_vmware.api [None req-2df5001e-3daa-4d21-82f3-a5fde08b46fb tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1118336, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.662880] env[63531]: DEBUG nova.network.neutron [req-4adebe96-9d08-4dfb-aa3c-33e1f0f4460c req-01cecca0-b420-4eb6-a9dd-80366ab5ade2 service nova] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 983.751140] env[63531]: DEBUG nova.network.neutron [req-4adebe96-9d08-4dfb-aa3c-33e1f0f4460c req-01cecca0-b420-4eb6-a9dd-80366ab5ade2 service nova] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.762774] env[63531]: DEBUG oslo_vmware.api [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118332, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.866706] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.865s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.867346] env[63531]: DEBUG nova.compute.manager [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 983.870648] env[63531]: DEBUG oslo_concurrency.lockutils [None req-795646f8-c711-4669-82d5-9d0c4ed80a2a tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.763s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.870879] env[63531]: DEBUG nova.objects.instance [None req-795646f8-c711-4669-82d5-9d0c4ed80a2a tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Lazy-loading 'resources' on Instance uuid 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 983.903908] env[63531]: DEBUG oslo_concurrency.lockutils [req-b6491d24-c054-4139-8eb2-456c3e367c39 req-886a2c1b-f1e7-4dd5-899b-288561189a0c service nova] Releasing lock "refresh_cache-3bf08a09-b6e6-4a58-8d35-7061a8e59afd" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.969543] env[63531]: DEBUG oslo_vmware.api [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118330, 'name': PowerOnVM_Task, 'duration_secs': 0.517669} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.969877] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 983.970170] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-0e3340c6-dc7e-4861-998e-02f4352287d7 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Updating instance '29dd6de3-2765-408c-acc0-da47e5e0a977' progress to 100 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 983.988289] env[63531]: DEBUG oslo_vmware.api [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]524e6de2-9d8a-1a8a-ed22-0d2472dad07f, 'name': SearchDatastore_Task, 'duration_secs': 0.041405} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.988667] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.988873] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3bf08a09-b6e6-4a58-8d35-7061a8e59afd] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 983.989154] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 983.989388] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.989560] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 983.989853] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8ae2a732-68a1-4071-a5ba-addd13c56c86 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.998185] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 983.998381] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 983.999183] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f5f93f7c-f6b0-400d-8d62-b201e8448f22 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.004653] env[63531]: DEBUG oslo_vmware.api [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 984.004653] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52fce8ef-cef6-8adc-0bb3-4c58c3f947f4" [ 984.004653] env[63531]: _type = "Task" [ 984.004653] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.015384] env[63531]: DEBUG oslo_vmware.api [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52fce8ef-cef6-8adc-0bb3-4c58c3f947f4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.099725] env[63531]: DEBUG oslo_vmware.api [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118334, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.152694] env[63531]: DEBUG oslo_vmware.api [None req-fc3f1fb2-bd92-4abe-bcb8-7277fece7b3a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118335, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.420391} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.153079] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc3f1fb2-bd92-4abe-bcb8-7277fece7b3a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 984.153366] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-fc3f1fb2-bd92-4abe-bcb8-7277fece7b3a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 984.153635] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-fc3f1fb2-bd92-4abe-bcb8-7277fece7b3a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 984.153896] env[63531]: INFO nova.compute.manager [None req-fc3f1fb2-bd92-4abe-bcb8-7277fece7b3a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Took 2.62 seconds to destroy the instance on the hypervisor. [ 984.154268] env[63531]: DEBUG oslo.service.loopingcall [None req-fc3f1fb2-bd92-4abe-bcb8-7277fece7b3a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 984.158721] env[63531]: DEBUG nova.compute.manager [-] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 984.158869] env[63531]: DEBUG nova.network.neutron [-] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 984.161444] env[63531]: DEBUG oslo_vmware.api [None req-2df5001e-3daa-4d21-82f3-a5fde08b46fb tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Task: {'id': task-1118336, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.348306} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.161774] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-2df5001e-3daa-4d21-82f3-a5fde08b46fb tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 984.162056] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-2df5001e-3daa-4d21-82f3-a5fde08b46fb tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 984.162331] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-2df5001e-3daa-4d21-82f3-a5fde08b46fb tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 984.162590] env[63531]: INFO nova.compute.manager [None req-2df5001e-3daa-4d21-82f3-a5fde08b46fb tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Took 1.95 seconds to destroy the instance on the hypervisor. [ 984.162939] env[63531]: DEBUG oslo.service.loopingcall [None req-2df5001e-3daa-4d21-82f3-a5fde08b46fb tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 984.163609] env[63531]: DEBUG nova.compute.manager [-] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 984.163765] env[63531]: DEBUG nova.network.neutron [-] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 984.256876] env[63531]: DEBUG oslo_concurrency.lockutils [req-4adebe96-9d08-4dfb-aa3c-33e1f0f4460c req-01cecca0-b420-4eb6-a9dd-80366ab5ade2 service nova] Releasing lock "refresh_cache-4757625a-8398-4b2d-b6fe-40eea8913068" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 984.257136] env[63531]: DEBUG oslo_concurrency.lockutils [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquired lock "refresh_cache-4757625a-8398-4b2d-b6fe-40eea8913068" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 984.257317] env[63531]: DEBUG nova.network.neutron [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 984.263792] env[63531]: DEBUG oslo_vmware.api [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118332, 'name': ReconfigVM_Task, 'duration_secs': 0.698378} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.264291] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: baf2db4e-2cdb-47f5-9a5c-97a233a459c9] Reconfigured VM instance instance-00000058 to attach disk [datastore2] baf2db4e-2cdb-47f5-9a5c-97a233a459c9/baf2db4e-2cdb-47f5-9a5c-97a233a459c9.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 984.264923] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b8601e61-475c-4cb6-9c2b-ffbe920094a4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.272549] env[63531]: DEBUG oslo_vmware.api [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Waiting for the task: (returnval){ [ 984.272549] env[63531]: value = "task-1118337" [ 984.272549] env[63531]: _type = "Task" [ 984.272549] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.281853] env[63531]: DEBUG oslo_vmware.api [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118337, 'name': Rename_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.374189] env[63531]: DEBUG nova.compute.utils [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 984.381333] env[63531]: DEBUG nova.compute.manager [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 984.381333] env[63531]: DEBUG nova.network.neutron [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 984.518049] env[63531]: DEBUG oslo_vmware.api [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52fce8ef-cef6-8adc-0bb3-4c58c3f947f4, 'name': SearchDatastore_Task, 'duration_secs': 0.009798} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.518896] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-28f05686-2d5c-4391-9ca3-55a65ab59ce1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.524211] env[63531]: DEBUG oslo_vmware.api [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 984.524211] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]529355ea-bd26-cc3a-1b25-20fdd5715e09" [ 984.524211] env[63531]: _type = "Task" [ 984.524211] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.534298] env[63531]: DEBUG oslo_vmware.api [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]529355ea-bd26-cc3a-1b25-20fdd5715e09, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.547581] env[63531]: DEBUG nova.policy [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '36cb2f0de54f417384fa1fb2210de4e4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8ba8b284854242c392aec5326e996239', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 984.596748] env[63531]: DEBUG nova.network.neutron [None req-ac717b99-ffee-40ff-8c50-9c164bf5d8ae tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Updating instance_info_cache with network_info: [{"id": "7433ce46-ff5a-4e78-bcc9-9dcf2264a01b", "address": "fa:16:3e:ad:77:b5", "network": {"id": "913045d9-e90b-4366-987c-7706b73a6147", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1696510389-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f71c7db0a3614995becc25d9a8dce787", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c297fe21-cd0b-4226-813b-a65d2358d034", "external-id": "nsx-vlan-transportzone-98", "segmentation_id": 98, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7433ce46-ff", "ovs_interfaceid": "7433ce46-ff5a-4e78-bcc9-9dcf2264a01b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 984.604924] env[63531]: DEBUG oslo_vmware.api [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118334, 'name': PowerOnVM_Task, 'duration_secs': 0.640688} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.605060] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: 0fb3b687-62cc-4296-b4ed-537aba18c880] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 984.606050] env[63531]: INFO nova.compute.manager [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: 0fb3b687-62cc-4296-b4ed-537aba18c880] Took 9.84 seconds to spawn the instance on the hypervisor. [ 984.606050] env[63531]: DEBUG nova.compute.manager [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: 0fb3b687-62cc-4296-b4ed-537aba18c880] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 984.606208] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf63cce9-d253-4d29-9c51-2a27e2a7f9b3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.676451] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af69042f-6175-4f42-bb15-b74fc1d0af0a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.684058] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f4509fe-817b-4e28-87ba-d2870328ad7e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.718098] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5559525d-d3d8-4f3f-b0d2-0d260bccf364 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.726031] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc2ade40-80c4-4fe9-b66b-b2c28ce157ae {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.739033] env[63531]: DEBUG nova.compute.provider_tree [None req-795646f8-c711-4669-82d5-9d0c4ed80a2a tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 984.786657] env[63531]: DEBUG oslo_vmware.api [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118337, 'name': Rename_Task, 'duration_secs': 0.259373} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.787045] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: baf2db4e-2cdb-47f5-9a5c-97a233a459c9] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 984.787300] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a4af7322-1325-47c6-9b0d-9b30f825f206 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.794034] env[63531]: DEBUG oslo_vmware.api [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Waiting for the task: (returnval){ [ 984.794034] env[63531]: value = "task-1118338" [ 984.794034] env[63531]: _type = "Task" [ 984.794034] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.794663] env[63531]: DEBUG nova.network.neutron [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 984.804168] env[63531]: DEBUG oslo_vmware.api [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118338, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.879493] env[63531]: DEBUG nova.compute.manager [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 984.944099] env[63531]: DEBUG nova.network.neutron [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Updating instance_info_cache with network_info: [{"id": "b37a092f-e03c-417f-bac1-3b2ebb5253d1", "address": "fa:16:3e:1e:cf:08", "network": {"id": "95ed2c7d-3d1a-4736-8e90-104002f5c40b", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-73549344-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9056486e3b640d0bcdb9369ed5d67ad", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04e15990-16e1-4cb2-b0f0-06c362e68c5e", "external-id": "nsx-vlan-transportzone-555", "segmentation_id": 555, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb37a092f-e0", "ovs_interfaceid": "b37a092f-e03c-417f-bac1-3b2ebb5253d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 985.036865] env[63531]: DEBUG oslo_vmware.api [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]529355ea-bd26-cc3a-1b25-20fdd5715e09, 'name': SearchDatastore_Task, 'duration_secs': 0.017914} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.037169] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 985.037438] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] 3bf08a09-b6e6-4a58-8d35-7061a8e59afd/3bf08a09-b6e6-4a58-8d35-7061a8e59afd.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 985.037712] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8d25bb8d-aebb-44b5-9c5f-6efbd62243d9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.044808] env[63531]: DEBUG oslo_vmware.api [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 985.044808] env[63531]: value = "task-1118339" [ 985.044808] env[63531]: _type = "Task" [ 985.044808] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.051649] env[63531]: DEBUG nova.network.neutron [-] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 985.056803] env[63531]: DEBUG oslo_vmware.api [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118339, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.099851] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ac717b99-ffee-40ff-8c50-9c164bf5d8ae tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Releasing lock "refresh_cache-ccf00b09-29fe-4a76-a8af-97774acd77de" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 985.124363] env[63531]: DEBUG nova.network.neutron [-] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 985.129574] env[63531]: INFO nova.compute.manager [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: 0fb3b687-62cc-4296-b4ed-537aba18c880] Took 23.20 seconds to build instance. [ 985.242185] env[63531]: DEBUG nova.scheduler.client.report [None req-795646f8-c711-4669-82d5-9d0c4ed80a2a tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 985.274867] env[63531]: DEBUG nova.compute.manager [req-4738dd48-580f-4b80-b2cc-7cd0176623f2 req-351cbc0c-d417-42ad-9383-5f8a5bd39423 service nova] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Received event network-vif-deleted-1bb5f5b0-d6c9-4b8d-be14-e6e98f2aeef1 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 985.274867] env[63531]: DEBUG nova.compute.manager [req-4738dd48-580f-4b80-b2cc-7cd0176623f2 req-351cbc0c-d417-42ad-9383-5f8a5bd39423 service nova] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Received event network-vif-deleted-a70530fb-7805-40a0-bea5-a2a666dc679d {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 985.310041] env[63531]: DEBUG oslo_vmware.api [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118338, 'name': PowerOnVM_Task, 'duration_secs': 0.464572} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.310414] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: baf2db4e-2cdb-47f5-9a5c-97a233a459c9] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 985.310648] env[63531]: INFO nova.compute.manager [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: baf2db4e-2cdb-47f5-9a5c-97a233a459c9] Took 8.18 seconds to spawn the instance on the hypervisor. [ 985.310823] env[63531]: DEBUG nova.compute.manager [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: baf2db4e-2cdb-47f5-9a5c-97a233a459c9] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 985.311677] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4def6701-07d3-466d-93d5-3e26289aa62f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.325034] env[63531]: DEBUG nova.network.neutron [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Successfully created port: a926b26b-b99b-49f2-b7dd-6ca158ead545 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 985.449637] env[63531]: DEBUG oslo_concurrency.lockutils [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Releasing lock "refresh_cache-4757625a-8398-4b2d-b6fe-40eea8913068" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 985.450090] env[63531]: DEBUG nova.compute.manager [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Instance network_info: |[{"id": "b37a092f-e03c-417f-bac1-3b2ebb5253d1", "address": "fa:16:3e:1e:cf:08", "network": {"id": "95ed2c7d-3d1a-4736-8e90-104002f5c40b", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-73549344-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9056486e3b640d0bcdb9369ed5d67ad", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04e15990-16e1-4cb2-b0f0-06c362e68c5e", "external-id": "nsx-vlan-transportzone-555", "segmentation_id": 555, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb37a092f-e0", "ovs_interfaceid": "b37a092f-e03c-417f-bac1-3b2ebb5253d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 985.450511] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1e:cf:08', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '04e15990-16e1-4cb2-b0f0-06c362e68c5e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b37a092f-e03c-417f-bac1-3b2ebb5253d1', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 985.459726] env[63531]: DEBUG oslo.service.loopingcall [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 985.460423] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 985.460685] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d5a29e03-bffb-4372-b704-6386f521b3f1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.495169] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 985.495169] env[63531]: value = "task-1118340" [ 985.495169] env[63531]: _type = "Task" [ 985.495169] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.509528] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118340, 'name': CreateVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.554635] env[63531]: DEBUG oslo_vmware.api [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118339, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.492746} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.555271] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] 3bf08a09-b6e6-4a58-8d35-7061a8e59afd/3bf08a09-b6e6-4a58-8d35-7061a8e59afd.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 985.555508] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3bf08a09-b6e6-4a58-8d35-7061a8e59afd] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 985.555813] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-15374ee9-b804-4e60-8f7b-521c51709f1e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.558010] env[63531]: INFO nova.compute.manager [-] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Took 1.40 seconds to deallocate network for instance. [ 985.565072] env[63531]: DEBUG oslo_vmware.api [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 985.565072] env[63531]: value = "task-1118341" [ 985.565072] env[63531]: _type = "Task" [ 985.565072] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.580752] env[63531]: DEBUG oslo_vmware.api [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118341, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.629941] env[63531]: INFO nova.compute.manager [-] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Took 1.47 seconds to deallocate network for instance. [ 985.632796] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Lock "0fb3b687-62cc-4296-b4ed-537aba18c880" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.705s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.643728] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac717b99-ffee-40ff-8c50-9c164bf5d8ae tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 985.643728] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bf22176a-1386-4340-a5e4-6b79406b2468 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.650028] env[63531]: DEBUG oslo_vmware.api [None req-ac717b99-ffee-40ff-8c50-9c164bf5d8ae tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 985.650028] env[63531]: value = "task-1118342" [ 985.650028] env[63531]: _type = "Task" [ 985.650028] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.658424] env[63531]: DEBUG oslo_vmware.api [None req-ac717b99-ffee-40ff-8c50-9c164bf5d8ae tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118342, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.748420] env[63531]: DEBUG oslo_concurrency.lockutils [None req-795646f8-c711-4669-82d5-9d0c4ed80a2a tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.878s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.752928] env[63531]: DEBUG oslo_concurrency.lockutils [None req-17a8f660-d313-4f51-a9dd-8d78a43ce44c tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 6.318s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.781232] env[63531]: INFO nova.scheduler.client.report [None req-795646f8-c711-4669-82d5-9d0c4ed80a2a tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Deleted allocations for instance 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f [ 985.837288] env[63531]: INFO nova.compute.manager [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: baf2db4e-2cdb-47f5-9a5c-97a233a459c9] Took 23.85 seconds to build instance. [ 985.893637] env[63531]: DEBUG nova.compute.manager [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 985.918145] env[63531]: DEBUG nova.virt.hardware [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 985.918478] env[63531]: DEBUG nova.virt.hardware [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 985.918649] env[63531]: DEBUG nova.virt.hardware [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 985.919420] env[63531]: DEBUG nova.virt.hardware [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 985.919420] env[63531]: DEBUG nova.virt.hardware [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 985.919420] env[63531]: DEBUG nova.virt.hardware [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 985.919585] env[63531]: DEBUG nova.virt.hardware [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 985.919630] env[63531]: DEBUG nova.virt.hardware [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 985.919785] env[63531]: DEBUG nova.virt.hardware [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 985.919954] env[63531]: DEBUG nova.virt.hardware [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 985.920157] env[63531]: DEBUG nova.virt.hardware [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 985.922174] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac6eaa9d-a327-4117-a0bd-99d57a8a0971 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.932911] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eeddba4c-340a-4a16-90a4-a19992c06258 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.007328] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118340, 'name': CreateVM_Task, 'duration_secs': 0.434049} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.007514] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 986.008262] env[63531]: DEBUG oslo_concurrency.lockutils [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.008431] env[63531]: DEBUG oslo_concurrency.lockutils [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.008763] env[63531]: DEBUG oslo_concurrency.lockutils [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 986.009154] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-afd7611d-5e40-4900-b849-4b824b5e6a80 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.016466] env[63531]: DEBUG oslo_vmware.api [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for the task: (returnval){ [ 986.016466] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52455881-173a-1806-9c33-7a3276594411" [ 986.016466] env[63531]: _type = "Task" [ 986.016466] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.021897] env[63531]: DEBUG oslo_vmware.api [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52455881-173a-1806-9c33-7a3276594411, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.077015] env[63531]: DEBUG oslo_vmware.api [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118341, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.088526} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.077015] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3bf08a09-b6e6-4a58-8d35-7061a8e59afd] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 986.077015] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6c8fdb4-9f23-4a5a-bfb4-4f9f34f7b98c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.102392] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3bf08a09-b6e6-4a58-8d35-7061a8e59afd] Reconfiguring VM instance instance-00000059 to attach disk [datastore2] 3bf08a09-b6e6-4a58-8d35-7061a8e59afd/3bf08a09-b6e6-4a58-8d35-7061a8e59afd.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 986.102720] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-27ea7178-3152-418f-b380-046b3be551c0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.121556] env[63531]: INFO nova.compute.manager [None req-fc3f1fb2-bd92-4abe-bcb8-7277fece7b3a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Took 0.56 seconds to detach 1 volumes for instance. [ 986.128460] env[63531]: DEBUG oslo_vmware.api [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 986.128460] env[63531]: value = "task-1118343" [ 986.128460] env[63531]: _type = "Task" [ 986.128460] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.136019] env[63531]: DEBUG oslo_vmware.api [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118343, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.138081] env[63531]: DEBUG oslo_concurrency.lockutils [None req-2df5001e-3daa-4d21-82f3-a5fde08b46fb tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.157905] env[63531]: DEBUG oslo_vmware.api [None req-ac717b99-ffee-40ff-8c50-9c164bf5d8ae tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118342, 'name': PowerOffVM_Task, 'duration_secs': 0.227114} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.158220] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac717b99-ffee-40ff-8c50-9c164bf5d8ae tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 986.159011] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39f9dc71-ae3c-404d-bf37-77608f7037fb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.180948] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ec99e56-d86b-4015-bbae-19aa08844e7e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.216517] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac717b99-ffee-40ff-8c50-9c164bf5d8ae tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 986.216855] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c6d31883-0b10-4b36-8147-20f60c30d7a6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.223826] env[63531]: DEBUG oslo_vmware.api [None req-ac717b99-ffee-40ff-8c50-9c164bf5d8ae tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 986.223826] env[63531]: value = "task-1118344" [ 986.223826] env[63531]: _type = "Task" [ 986.223826] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.234431] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac717b99-ffee-40ff-8c50-9c164bf5d8ae tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] VM already powered off {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 986.234663] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-ac717b99-ffee-40ff-8c50-9c164bf5d8ae tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 986.234918] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ac717b99-ffee-40ff-8c50-9c164bf5d8ae tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.235090] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ac717b99-ffee-40ff-8c50-9c164bf5d8ae tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.235278] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac717b99-ffee-40ff-8c50-9c164bf5d8ae tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 986.235529] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-03154e6c-e422-4f89-aad1-a21cc542089d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.246624] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac717b99-ffee-40ff-8c50-9c164bf5d8ae tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 986.246874] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-ac717b99-ffee-40ff-8c50-9c164bf5d8ae tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 986.247549] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b0fd5832-cd34-478f-8c9e-0f249e488b94 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.253143] env[63531]: DEBUG oslo_vmware.api [None req-ac717b99-ffee-40ff-8c50-9c164bf5d8ae tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 986.253143] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52ecae1f-18b2-af29-a377-d2c5479ff1fb" [ 986.253143] env[63531]: _type = "Task" [ 986.253143] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.263328] env[63531]: DEBUG oslo_vmware.api [None req-ac717b99-ffee-40ff-8c50-9c164bf5d8ae tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52ecae1f-18b2-af29-a377-d2c5479ff1fb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.287923] env[63531]: DEBUG oslo_concurrency.lockutils [None req-795646f8-c711-4669-82d5-9d0c4ed80a2a tempest-AttachVolumeShelveTestJSON-88751117 tempest-AttachVolumeShelveTestJSON-88751117-project-member] Lock "2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.436s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.341902] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0fcb054-54cf-4ee1-845f-b4ea50148b14 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Lock "baf2db4e-2cdb-47f5-9a5c-97a233a459c9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.370s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.407914] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1b6ffe13-6d20-4c93-b882-81d0eee2d420 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "29dd6de3-2765-408c-acc0-da47e5e0a977" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.407914] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1b6ffe13-6d20-4c93-b882-81d0eee2d420 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "29dd6de3-2765-408c-acc0-da47e5e0a977" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.408098] env[63531]: DEBUG nova.compute.manager [None req-1b6ffe13-6d20-4c93-b882-81d0eee2d420 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Going to confirm migration 4 {{(pid=63531) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 986.471381] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Acquiring lock "b7b2b7ac-2321-4cf3-b2ed-4877b893af63" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.471633] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lock "b7b2b7ac-2321-4cf3-b2ed-4877b893af63" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.524057] env[63531]: DEBUG oslo_vmware.api [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52455881-173a-1806-9c33-7a3276594411, 'name': SearchDatastore_Task, 'duration_secs': 0.067496} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.525205] env[63531]: DEBUG oslo_concurrency.lockutils [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.525450] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 986.525749] env[63531]: DEBUG oslo_concurrency.lockutils [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.525927] env[63531]: DEBUG oslo_concurrency.lockutils [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.526133] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 986.526883] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa1876a6-4138-4c7d-83fd-51ea0e82e55b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.529379] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-34de29e1-da28-49c1-abd6-fcfc955a7aa4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.535641] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe95fd0c-a5fe-4f86-9c0c-a3d4b36b3417 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.567512] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2993fe5-5cf0-4ab0-add3-ce5b4c1bac3e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.570227] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 986.570419] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 986.571142] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7c3484c6-c18b-45ff-a3e5-28db66dd3182 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.579585] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fba8d259-9582-49c2-be81-ab25d6bfeea9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.583626] env[63531]: DEBUG oslo_vmware.api [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for the task: (returnval){ [ 986.583626] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52d232bc-4ce1-4fe7-ad95-aa9aef58d750" [ 986.583626] env[63531]: _type = "Task" [ 986.583626] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.595196] env[63531]: DEBUG nova.compute.provider_tree [None req-17a8f660-d313-4f51-a9dd-8d78a43ce44c tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 986.601675] env[63531]: DEBUG oslo_vmware.api [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52d232bc-4ce1-4fe7-ad95-aa9aef58d750, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.628353] env[63531]: DEBUG oslo_concurrency.lockutils [None req-fc3f1fb2-bd92-4abe-bcb8-7277fece7b3a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.639405] env[63531]: DEBUG oslo_vmware.api [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118343, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.764244] env[63531]: DEBUG oslo_vmware.api [None req-ac717b99-ffee-40ff-8c50-9c164bf5d8ae tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52ecae1f-18b2-af29-a377-d2c5479ff1fb, 'name': SearchDatastore_Task, 'duration_secs': 0.077328} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.765073] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-695a5645-d059-47fb-b841-f2256df93897 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.770374] env[63531]: DEBUG oslo_vmware.api [None req-ac717b99-ffee-40ff-8c50-9c164bf5d8ae tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 986.770374] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]528b0954-1604-422f-5c6e-aa27e5cfc1be" [ 986.770374] env[63531]: _type = "Task" [ 986.770374] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.778945] env[63531]: DEBUG oslo_vmware.api [None req-ac717b99-ffee-40ff-8c50-9c164bf5d8ae tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]528b0954-1604-422f-5c6e-aa27e5cfc1be, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.974246] env[63531]: DEBUG nova.compute.manager [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 986.983646] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1b6ffe13-6d20-4c93-b882-81d0eee2d420 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "refresh_cache-29dd6de3-2765-408c-acc0-da47e5e0a977" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.983646] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1b6ffe13-6d20-4c93-b882-81d0eee2d420 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquired lock "refresh_cache-29dd6de3-2765-408c-acc0-da47e5e0a977" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.983646] env[63531]: DEBUG nova.network.neutron [None req-1b6ffe13-6d20-4c93-b882-81d0eee2d420 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 986.983883] env[63531]: DEBUG nova.objects.instance [None req-1b6ffe13-6d20-4c93-b882-81d0eee2d420 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lazy-loading 'info_cache' on Instance uuid 29dd6de3-2765-408c-acc0-da47e5e0a977 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 987.094268] env[63531]: DEBUG oslo_vmware.api [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52d232bc-4ce1-4fe7-ad95-aa9aef58d750, 'name': SearchDatastore_Task, 'duration_secs': 0.044127} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.095039] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4db8b65c-ee08-47e7-bc62-3f2a13f7783a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.097981] env[63531]: DEBUG nova.scheduler.client.report [None req-17a8f660-d313-4f51-a9dd-8d78a43ce44c tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 987.105820] env[63531]: DEBUG oslo_vmware.api [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for the task: (returnval){ [ 987.105820] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]522d8df1-57f5-f15b-85e4-701ed34a5491" [ 987.105820] env[63531]: _type = "Task" [ 987.105820] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.113283] env[63531]: DEBUG oslo_vmware.api [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]522d8df1-57f5-f15b-85e4-701ed34a5491, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.138361] env[63531]: DEBUG oslo_vmware.api [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118343, 'name': ReconfigVM_Task, 'duration_secs': 0.91051} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.138655] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3bf08a09-b6e6-4a58-8d35-7061a8e59afd] Reconfigured VM instance instance-00000059 to attach disk [datastore2] 3bf08a09-b6e6-4a58-8d35-7061a8e59afd/3bf08a09-b6e6-4a58-8d35-7061a8e59afd.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 987.139345] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1ea2f1ed-c9a1-4d6e-8620-32aa54a5e3f8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.145863] env[63531]: DEBUG oslo_vmware.api [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 987.145863] env[63531]: value = "task-1118345" [ 987.145863] env[63531]: _type = "Task" [ 987.145863] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.154883] env[63531]: DEBUG oslo_vmware.api [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118345, 'name': Rename_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.162569] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9b48ef40-e853-4488-bc73-2ca4984fb08c tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Acquiring lock "0fb3b687-62cc-4296-b4ed-537aba18c880" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.162915] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9b48ef40-e853-4488-bc73-2ca4984fb08c tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Lock "0fb3b687-62cc-4296-b4ed-537aba18c880" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.163189] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9b48ef40-e853-4488-bc73-2ca4984fb08c tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Acquiring lock "0fb3b687-62cc-4296-b4ed-537aba18c880-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.163421] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9b48ef40-e853-4488-bc73-2ca4984fb08c tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Lock "0fb3b687-62cc-4296-b4ed-537aba18c880-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.163847] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9b48ef40-e853-4488-bc73-2ca4984fb08c tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Lock "0fb3b687-62cc-4296-b4ed-537aba18c880-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.169649] env[63531]: INFO nova.compute.manager [None req-9b48ef40-e853-4488-bc73-2ca4984fb08c tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: 0fb3b687-62cc-4296-b4ed-537aba18c880] Terminating instance [ 987.170158] env[63531]: DEBUG nova.compute.manager [None req-9b48ef40-e853-4488-bc73-2ca4984fb08c tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: 0fb3b687-62cc-4296-b4ed-537aba18c880] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 987.170384] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-9b48ef40-e853-4488-bc73-2ca4984fb08c tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: 0fb3b687-62cc-4296-b4ed-537aba18c880] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 987.171358] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de270dda-ee14-47e4-9537-688e3d934ac9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.179701] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b48ef40-e853-4488-bc73-2ca4984fb08c tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: 0fb3b687-62cc-4296-b4ed-537aba18c880] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 987.179834] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-75eb4e5a-8c80-4255-8a47-b4c1b95ed624 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.186643] env[63531]: DEBUG oslo_vmware.api [None req-9b48ef40-e853-4488-bc73-2ca4984fb08c tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Waiting for the task: (returnval){ [ 987.186643] env[63531]: value = "task-1118346" [ 987.186643] env[63531]: _type = "Task" [ 987.186643] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.194873] env[63531]: DEBUG oslo_vmware.api [None req-9b48ef40-e853-4488-bc73-2ca4984fb08c tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118346, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.196990] env[63531]: DEBUG nova.compute.manager [req-43cc50b7-8001-4443-889f-17a809770438 req-bd247dc4-f260-4906-895e-eedd5226f645 service nova] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Received event network-vif-plugged-a926b26b-b99b-49f2-b7dd-6ca158ead545 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 987.197248] env[63531]: DEBUG oslo_concurrency.lockutils [req-43cc50b7-8001-4443-889f-17a809770438 req-bd247dc4-f260-4906-895e-eedd5226f645 service nova] Acquiring lock "e742d3af-5a45-4274-adfb-c6138bf84d6a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.197462] env[63531]: DEBUG oslo_concurrency.lockutils [req-43cc50b7-8001-4443-889f-17a809770438 req-bd247dc4-f260-4906-895e-eedd5226f645 service nova] Lock "e742d3af-5a45-4274-adfb-c6138bf84d6a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.197635] env[63531]: DEBUG oslo_concurrency.lockutils [req-43cc50b7-8001-4443-889f-17a809770438 req-bd247dc4-f260-4906-895e-eedd5226f645 service nova] Lock "e742d3af-5a45-4274-adfb-c6138bf84d6a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.197845] env[63531]: DEBUG nova.compute.manager [req-43cc50b7-8001-4443-889f-17a809770438 req-bd247dc4-f260-4906-895e-eedd5226f645 service nova] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] No waiting events found dispatching network-vif-plugged-a926b26b-b99b-49f2-b7dd-6ca158ead545 {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 987.198027] env[63531]: WARNING nova.compute.manager [req-43cc50b7-8001-4443-889f-17a809770438 req-bd247dc4-f260-4906-895e-eedd5226f645 service nova] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Received unexpected event network-vif-plugged-a926b26b-b99b-49f2-b7dd-6ca158ead545 for instance with vm_state building and task_state spawning. [ 987.229275] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0f3b984b-d279-4944-8a9e-cb9bdebbf475 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Acquiring lock "baf2db4e-2cdb-47f5-9a5c-97a233a459c9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.229549] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0f3b984b-d279-4944-8a9e-cb9bdebbf475 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Lock "baf2db4e-2cdb-47f5-9a5c-97a233a459c9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.229769] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0f3b984b-d279-4944-8a9e-cb9bdebbf475 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Acquiring lock "baf2db4e-2cdb-47f5-9a5c-97a233a459c9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.230046] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0f3b984b-d279-4944-8a9e-cb9bdebbf475 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Lock "baf2db4e-2cdb-47f5-9a5c-97a233a459c9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.230239] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0f3b984b-d279-4944-8a9e-cb9bdebbf475 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Lock "baf2db4e-2cdb-47f5-9a5c-97a233a459c9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.232420] env[63531]: INFO nova.compute.manager [None req-0f3b984b-d279-4944-8a9e-cb9bdebbf475 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: baf2db4e-2cdb-47f5-9a5c-97a233a459c9] Terminating instance [ 987.234806] env[63531]: DEBUG nova.compute.manager [None req-0f3b984b-d279-4944-8a9e-cb9bdebbf475 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: baf2db4e-2cdb-47f5-9a5c-97a233a459c9] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 987.235013] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-0f3b984b-d279-4944-8a9e-cb9bdebbf475 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: baf2db4e-2cdb-47f5-9a5c-97a233a459c9] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 987.235912] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29892b13-3db2-4952-b7ab-6ac5cc3b3707 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.245450] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f3b984b-d279-4944-8a9e-cb9bdebbf475 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: baf2db4e-2cdb-47f5-9a5c-97a233a459c9] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 987.245715] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-00e794a1-7e9a-48aa-823f-8826da00d431 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.252130] env[63531]: DEBUG oslo_vmware.api [None req-0f3b984b-d279-4944-8a9e-cb9bdebbf475 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Waiting for the task: (returnval){ [ 987.252130] env[63531]: value = "task-1118347" [ 987.252130] env[63531]: _type = "Task" [ 987.252130] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.261277] env[63531]: DEBUG oslo_vmware.api [None req-0f3b984b-d279-4944-8a9e-cb9bdebbf475 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118347, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.281734] env[63531]: DEBUG oslo_vmware.api [None req-ac717b99-ffee-40ff-8c50-9c164bf5d8ae tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]528b0954-1604-422f-5c6e-aa27e5cfc1be, 'name': SearchDatastore_Task, 'duration_secs': 0.026862} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.281910] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ac717b99-ffee-40ff-8c50-9c164bf5d8ae tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.282203] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac717b99-ffee-40ff-8c50-9c164bf5d8ae tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] ccf00b09-29fe-4a76-a8af-97774acd77de/9d550399-588c-4bd5-8436-eba9b523bd13-rescue.vmdk. {{(pid=63531) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 987.283681] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b15a683f-ba52-43d9-a863-c16ceebae91f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.290523] env[63531]: DEBUG oslo_vmware.api [None req-ac717b99-ffee-40ff-8c50-9c164bf5d8ae tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 987.290523] env[63531]: value = "task-1118348" [ 987.290523] env[63531]: _type = "Task" [ 987.290523] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.300594] env[63531]: DEBUG oslo_vmware.api [None req-ac717b99-ffee-40ff-8c50-9c164bf5d8ae tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118348, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.435700] env[63531]: DEBUG nova.network.neutron [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Successfully updated port: a926b26b-b99b-49f2-b7dd-6ca158ead545 {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 987.503954] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.622590] env[63531]: DEBUG oslo_vmware.api [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]522d8df1-57f5-f15b-85e4-701ed34a5491, 'name': SearchDatastore_Task, 'duration_secs': 0.010289} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.622942] env[63531]: DEBUG oslo_concurrency.lockutils [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.623237] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] 4757625a-8398-4b2d-b6fe-40eea8913068/4757625a-8398-4b2d-b6fe-40eea8913068.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 987.623525] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b50836ce-2aca-4c43-af98-9ac620758e8e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.631330] env[63531]: DEBUG oslo_vmware.api [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for the task: (returnval){ [ 987.631330] env[63531]: value = "task-1118349" [ 987.631330] env[63531]: _type = "Task" [ 987.631330] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.640222] env[63531]: DEBUG oslo_vmware.api [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118349, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.656657] env[63531]: DEBUG oslo_vmware.api [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118345, 'name': Rename_Task, 'duration_secs': 0.17883} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.656976] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3bf08a09-b6e6-4a58-8d35-7061a8e59afd] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 987.657314] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7180e566-62b3-4354-acfa-5dbb2d420856 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.663855] env[63531]: DEBUG oslo_vmware.api [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 987.663855] env[63531]: value = "task-1118350" [ 987.663855] env[63531]: _type = "Task" [ 987.663855] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.674755] env[63531]: DEBUG oslo_vmware.api [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118350, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.697871] env[63531]: DEBUG oslo_vmware.api [None req-9b48ef40-e853-4488-bc73-2ca4984fb08c tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118346, 'name': PowerOffVM_Task, 'duration_secs': 0.210624} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.698219] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b48ef40-e853-4488-bc73-2ca4984fb08c tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: 0fb3b687-62cc-4296-b4ed-537aba18c880] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 987.698413] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-9b48ef40-e853-4488-bc73-2ca4984fb08c tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: 0fb3b687-62cc-4296-b4ed-537aba18c880] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 987.698693] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-27e67aa0-912e-40bb-a4ce-3b15157ed216 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.767349] env[63531]: DEBUG oslo_vmware.api [None req-0f3b984b-d279-4944-8a9e-cb9bdebbf475 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118347, 'name': PowerOffVM_Task, 'duration_secs': 0.16172} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.767934] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f3b984b-d279-4944-8a9e-cb9bdebbf475 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: baf2db4e-2cdb-47f5-9a5c-97a233a459c9] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 987.768293] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-0f3b984b-d279-4944-8a9e-cb9bdebbf475 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: baf2db4e-2cdb-47f5-9a5c-97a233a459c9] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 987.768732] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-63fc92e4-a818-4160-ba8c-2f8118761683 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.798538] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-9b48ef40-e853-4488-bc73-2ca4984fb08c tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: 0fb3b687-62cc-4296-b4ed-537aba18c880] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 987.798889] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-9b48ef40-e853-4488-bc73-2ca4984fb08c tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: 0fb3b687-62cc-4296-b4ed-537aba18c880] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 987.799196] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b48ef40-e853-4488-bc73-2ca4984fb08c tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Deleting the datastore file [datastore1] 0fb3b687-62cc-4296-b4ed-537aba18c880 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 987.802450] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ada9fa88-4601-40c4-9a77-ec8c76e7024f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.804285] env[63531]: DEBUG oslo_vmware.api [None req-ac717b99-ffee-40ff-8c50-9c164bf5d8ae tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118348, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.504458} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.804529] env[63531]: INFO nova.virt.vmwareapi.ds_util [None req-ac717b99-ffee-40ff-8c50-9c164bf5d8ae tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] ccf00b09-29fe-4a76-a8af-97774acd77de/9d550399-588c-4bd5-8436-eba9b523bd13-rescue.vmdk. [ 987.805664] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a167feac-e01d-4ec9-b208-f54663002085 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.810827] env[63531]: DEBUG oslo_vmware.api [None req-9b48ef40-e853-4488-bc73-2ca4984fb08c tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Waiting for the task: (returnval){ [ 987.810827] env[63531]: value = "task-1118353" [ 987.810827] env[63531]: _type = "Task" [ 987.810827] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.841959] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-ac717b99-ffee-40ff-8c50-9c164bf5d8ae tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Reconfiguring VM instance instance-00000051 to attach disk [datastore1] ccf00b09-29fe-4a76-a8af-97774acd77de/9d550399-588c-4bd5-8436-eba9b523bd13-rescue.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 987.843751] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8940b0e9-dd6b-4d4f-8d91-86dfe890bd47 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.859379] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-0f3b984b-d279-4944-8a9e-cb9bdebbf475 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: baf2db4e-2cdb-47f5-9a5c-97a233a459c9] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 987.859651] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-0f3b984b-d279-4944-8a9e-cb9bdebbf475 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: baf2db4e-2cdb-47f5-9a5c-97a233a459c9] Deleting contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 987.859887] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f3b984b-d279-4944-8a9e-cb9bdebbf475 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Deleting the datastore file [datastore2] baf2db4e-2cdb-47f5-9a5c-97a233a459c9 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 987.863550] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d64bfad2-ee10-4a34-a739-8cf2b37eb935 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.865934] env[63531]: DEBUG oslo_vmware.api [None req-9b48ef40-e853-4488-bc73-2ca4984fb08c tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118353, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.871538] env[63531]: DEBUG oslo_vmware.api [None req-0f3b984b-d279-4944-8a9e-cb9bdebbf475 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Waiting for the task: (returnval){ [ 987.871538] env[63531]: value = "task-1118354" [ 987.871538] env[63531]: _type = "Task" [ 987.871538] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.873011] env[63531]: DEBUG oslo_vmware.api [None req-ac717b99-ffee-40ff-8c50-9c164bf5d8ae tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 987.873011] env[63531]: value = "task-1118355" [ 987.873011] env[63531]: _type = "Task" [ 987.873011] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.884185] env[63531]: DEBUG oslo_vmware.api [None req-0f3b984b-d279-4944-8a9e-cb9bdebbf475 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118354, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.887699] env[63531]: DEBUG oslo_vmware.api [None req-ac717b99-ffee-40ff-8c50-9c164bf5d8ae tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118355, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.937174] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "refresh_cache-e742d3af-5a45-4274-adfb-c6138bf84d6a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 987.937244] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquired lock "refresh_cache-e742d3af-5a45-4274-adfb-c6138bf84d6a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.937404] env[63531]: DEBUG nova.network.neutron [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 988.116069] env[63531]: DEBUG oslo_concurrency.lockutils [None req-17a8f660-d313-4f51-a9dd-8d78a43ce44c tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.364s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.119892] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.246s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.121915] env[63531]: INFO nova.compute.claims [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 988.144284] env[63531]: DEBUG oslo_vmware.api [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118349, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.181540] env[63531]: DEBUG oslo_vmware.api [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118350, 'name': PowerOnVM_Task} progress is 78%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.298929] env[63531]: DEBUG nova.network.neutron [None req-1b6ffe13-6d20-4c93-b882-81d0eee2d420 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Updating instance_info_cache with network_info: [{"id": "262abfb7-7bcf-40e6-9ce3-b502de45b3b4", "address": "fa:16:3e:0e:00:c9", "network": {"id": "0667c176-cfdf-4e43-b152-e00bdc224448", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-243715985-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.217", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "308eabafea9c4188a58a0f1c22074d2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8f580e6-1d86-41ee-9ebe-c531cb9299c6", "external-id": "nsx-vlan-transportzone-150", "segmentation_id": 150, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap262abfb7-7b", "ovs_interfaceid": "262abfb7-7bcf-40e6-9ce3-b502de45b3b4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 988.320965] env[63531]: DEBUG oslo_vmware.api [None req-9b48ef40-e853-4488-bc73-2ca4984fb08c tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118353, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.376936} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.321281] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b48ef40-e853-4488-bc73-2ca4984fb08c tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 988.321494] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-9b48ef40-e853-4488-bc73-2ca4984fb08c tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: 0fb3b687-62cc-4296-b4ed-537aba18c880] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 988.321770] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-9b48ef40-e853-4488-bc73-2ca4984fb08c tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: 0fb3b687-62cc-4296-b4ed-537aba18c880] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 988.321861] env[63531]: INFO nova.compute.manager [None req-9b48ef40-e853-4488-bc73-2ca4984fb08c tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: 0fb3b687-62cc-4296-b4ed-537aba18c880] Took 1.15 seconds to destroy the instance on the hypervisor. [ 988.322098] env[63531]: DEBUG oslo.service.loopingcall [None req-9b48ef40-e853-4488-bc73-2ca4984fb08c tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 988.322321] env[63531]: DEBUG nova.compute.manager [-] [instance: 0fb3b687-62cc-4296-b4ed-537aba18c880] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 988.322418] env[63531]: DEBUG nova.network.neutron [-] [instance: 0fb3b687-62cc-4296-b4ed-537aba18c880] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 988.387266] env[63531]: DEBUG oslo_vmware.api [None req-0f3b984b-d279-4944-8a9e-cb9bdebbf475 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118354, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.390420] env[63531]: DEBUG oslo_vmware.api [None req-ac717b99-ffee-40ff-8c50-9c164bf5d8ae tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118355, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.484810] env[63531]: DEBUG nova.network.neutron [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 988.641082] env[63531]: DEBUG oslo_vmware.api [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118349, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.663776} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.641356] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] 4757625a-8398-4b2d-b6fe-40eea8913068/4757625a-8398-4b2d-b6fe-40eea8913068.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 988.641576] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 988.641831] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-83da87cc-ea71-4a0e-8359-4bab81f6c0bc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.651237] env[63531]: DEBUG oslo_vmware.api [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for the task: (returnval){ [ 988.651237] env[63531]: value = "task-1118356" [ 988.651237] env[63531]: _type = "Task" [ 988.651237] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.662550] env[63531]: DEBUG oslo_vmware.api [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118356, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.675795] env[63531]: DEBUG oslo_vmware.api [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118350, 'name': PowerOnVM_Task, 'duration_secs': 0.886576} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.676347] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3bf08a09-b6e6-4a58-8d35-7061a8e59afd] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 988.676347] env[63531]: INFO nova.compute.manager [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3bf08a09-b6e6-4a58-8d35-7061a8e59afd] Took 9.01 seconds to spawn the instance on the hypervisor. [ 988.676600] env[63531]: DEBUG nova.compute.manager [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3bf08a09-b6e6-4a58-8d35-7061a8e59afd] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 988.677420] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efc7426a-f789-4c91-9975-99cea7b4e256 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.694184] env[63531]: INFO nova.scheduler.client.report [None req-17a8f660-d313-4f51-a9dd-8d78a43ce44c tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Deleted allocation for migration d5d6b5dc-c0a9-41c2-96a9-6aa3ec666dcd [ 988.709080] env[63531]: DEBUG nova.network.neutron [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Updating instance_info_cache with network_info: [{"id": "a926b26b-b99b-49f2-b7dd-6ca158ead545", "address": "fa:16:3e:99:7c:93", "network": {"id": "caf86428-59a5-4e50-9135-2ab8473af64d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1673994696-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ba8b284854242c392aec5326e996239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa926b26b-b9", "ovs_interfaceid": "a926b26b-b99b-49f2-b7dd-6ca158ead545", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 988.802397] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1b6ffe13-6d20-4c93-b882-81d0eee2d420 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Releasing lock "refresh_cache-29dd6de3-2765-408c-acc0-da47e5e0a977" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.802777] env[63531]: DEBUG nova.objects.instance [None req-1b6ffe13-6d20-4c93-b882-81d0eee2d420 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lazy-loading 'migration_context' on Instance uuid 29dd6de3-2765-408c-acc0-da47e5e0a977 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 988.885929] env[63531]: DEBUG oslo_vmware.api [None req-0f3b984b-d279-4944-8a9e-cb9bdebbf475 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118354, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.518113} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.886520] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f3b984b-d279-4944-8a9e-cb9bdebbf475 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 988.886718] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-0f3b984b-d279-4944-8a9e-cb9bdebbf475 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: baf2db4e-2cdb-47f5-9a5c-97a233a459c9] Deleted contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 988.886911] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-0f3b984b-d279-4944-8a9e-cb9bdebbf475 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: baf2db4e-2cdb-47f5-9a5c-97a233a459c9] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 988.887137] env[63531]: INFO nova.compute.manager [None req-0f3b984b-d279-4944-8a9e-cb9bdebbf475 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: baf2db4e-2cdb-47f5-9a5c-97a233a459c9] Took 1.65 seconds to destroy the instance on the hypervisor. [ 988.887382] env[63531]: DEBUG oslo.service.loopingcall [None req-0f3b984b-d279-4944-8a9e-cb9bdebbf475 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 988.887585] env[63531]: DEBUG nova.compute.manager [-] [instance: baf2db4e-2cdb-47f5-9a5c-97a233a459c9] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 988.887684] env[63531]: DEBUG nova.network.neutron [-] [instance: baf2db4e-2cdb-47f5-9a5c-97a233a459c9] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 988.891930] env[63531]: DEBUG oslo_vmware.api [None req-ac717b99-ffee-40ff-8c50-9c164bf5d8ae tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118355, 'name': ReconfigVM_Task, 'duration_secs': 0.71512} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.892465] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-ac717b99-ffee-40ff-8c50-9c164bf5d8ae tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Reconfigured VM instance instance-00000051 to attach disk [datastore1] ccf00b09-29fe-4a76-a8af-97774acd77de/9d550399-588c-4bd5-8436-eba9b523bd13-rescue.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 988.893308] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2b6cf3e-c83e-413a-98fd-0a452865877f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.920749] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-da18009d-881d-4c83-9e38-31bae7a88eba {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.938572] env[63531]: DEBUG oslo_vmware.api [None req-ac717b99-ffee-40ff-8c50-9c164bf5d8ae tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 988.938572] env[63531]: value = "task-1118357" [ 988.938572] env[63531]: _type = "Task" [ 988.938572] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.946987] env[63531]: DEBUG oslo_vmware.api [None req-ac717b99-ffee-40ff-8c50-9c164bf5d8ae tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118357, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.163610] env[63531]: DEBUG oslo_vmware.api [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118356, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063458} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.163945] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 989.164618] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16e47f21-5c36-442f-a3e3-94ef1372e426 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.189932] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Reconfiguring VM instance instance-0000005a to attach disk [datastore2] 4757625a-8398-4b2d-b6fe-40eea8913068/4757625a-8398-4b2d-b6fe-40eea8913068.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 989.206475] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-edd6cf97-6b36-44e4-8f7d-d23e67258392 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.219080] env[63531]: DEBUG oslo_concurrency.lockutils [None req-17a8f660-d313-4f51-a9dd-8d78a43ce44c tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "27c23b7c-a68f-43c9-a125-bd657feb3c5b" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 12.205s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.222135] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Releasing lock "refresh_cache-e742d3af-5a45-4274-adfb-c6138bf84d6a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.224390] env[63531]: DEBUG nova.compute.manager [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Instance network_info: |[{"id": "a926b26b-b99b-49f2-b7dd-6ca158ead545", "address": "fa:16:3e:99:7c:93", "network": {"id": "caf86428-59a5-4e50-9135-2ab8473af64d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1673994696-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ba8b284854242c392aec5326e996239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa926b26b-b9", "ovs_interfaceid": "a926b26b-b99b-49f2-b7dd-6ca158ead545", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 989.226808] env[63531]: INFO nova.compute.manager [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3bf08a09-b6e6-4a58-8d35-7061a8e59afd] Took 26.63 seconds to build instance. [ 989.226808] env[63531]: DEBUG nova.network.neutron [-] [instance: 0fb3b687-62cc-4296-b4ed-537aba18c880] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.227049] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:99:7c:93', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '91b0f7e5-0d1a-46e2-bf73-09656211dea2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a926b26b-b99b-49f2-b7dd-6ca158ead545', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 989.236400] env[63531]: DEBUG oslo.service.loopingcall [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 989.239725] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 989.241946] env[63531]: DEBUG nova.compute.manager [req-ec79c6d9-99c8-42df-9dd8-2f9063496433 req-b0460b96-b2e0-4388-8231-81cdb09adf8a service nova] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Received event network-changed-a926b26b-b99b-49f2-b7dd-6ca158ead545 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 989.241946] env[63531]: DEBUG nova.compute.manager [req-ec79c6d9-99c8-42df-9dd8-2f9063496433 req-b0460b96-b2e0-4388-8231-81cdb09adf8a service nova] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Refreshing instance network info cache due to event network-changed-a926b26b-b99b-49f2-b7dd-6ca158ead545. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 989.242100] env[63531]: DEBUG oslo_concurrency.lockutils [req-ec79c6d9-99c8-42df-9dd8-2f9063496433 req-b0460b96-b2e0-4388-8231-81cdb09adf8a service nova] Acquiring lock "refresh_cache-e742d3af-5a45-4274-adfb-c6138bf84d6a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.242155] env[63531]: DEBUG oslo_concurrency.lockutils [req-ec79c6d9-99c8-42df-9dd8-2f9063496433 req-b0460b96-b2e0-4388-8231-81cdb09adf8a service nova] Acquired lock "refresh_cache-e742d3af-5a45-4274-adfb-c6138bf84d6a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.242823] env[63531]: DEBUG nova.network.neutron [req-ec79c6d9-99c8-42df-9dd8-2f9063496433 req-b0460b96-b2e0-4388-8231-81cdb09adf8a service nova] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Refreshing network info cache for port a926b26b-b99b-49f2-b7dd-6ca158ead545 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 989.245991] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-60913ccf-3306-4bea-b6ec-8f91b9d159b2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.262352] env[63531]: DEBUG oslo_vmware.api [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for the task: (returnval){ [ 989.262352] env[63531]: value = "task-1118358" [ 989.262352] env[63531]: _type = "Task" [ 989.262352] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.273296] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 989.273296] env[63531]: value = "task-1118359" [ 989.273296] env[63531]: _type = "Task" [ 989.273296] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.282508] env[63531]: DEBUG oslo_vmware.api [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118358, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.290084] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118359, 'name': CreateVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.305681] env[63531]: DEBUG nova.objects.base [None req-1b6ffe13-6d20-4c93-b882-81d0eee2d420 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Object Instance<29dd6de3-2765-408c-acc0-da47e5e0a977> lazy-loaded attributes: info_cache,migration_context {{(pid=63531) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 989.306609] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e138fa5f-a00d-41dc-929a-afe35ecfa749 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.329943] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2866d48f-7d28-4a29-b510-8778ad709833 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.335987] env[63531]: DEBUG oslo_vmware.api [None req-1b6ffe13-6d20-4c93-b882-81d0eee2d420 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 989.335987] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52ff628f-1382-16ab-2c3d-71a8a65c35c7" [ 989.335987] env[63531]: _type = "Task" [ 989.335987] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.346450] env[63531]: DEBUG oslo_vmware.api [None req-1b6ffe13-6d20-4c93-b882-81d0eee2d420 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52ff628f-1382-16ab-2c3d-71a8a65c35c7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.454100] env[63531]: DEBUG oslo_vmware.api [None req-ac717b99-ffee-40ff-8c50-9c164bf5d8ae tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118357, 'name': ReconfigVM_Task, 'duration_secs': 0.456095} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.454100] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac717b99-ffee-40ff-8c50-9c164bf5d8ae tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 989.454100] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-68372162-8b7d-43d9-91c1-3087a9ce5ac2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.459753] env[63531]: DEBUG oslo_vmware.api [None req-ac717b99-ffee-40ff-8c50-9c164bf5d8ae tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 989.459753] env[63531]: value = "task-1118360" [ 989.459753] env[63531]: _type = "Task" [ 989.459753] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.474849] env[63531]: DEBUG oslo_vmware.api [None req-ac717b99-ffee-40ff-8c50-9c164bf5d8ae tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118360, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.570812] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10985d72-528e-4cde-bfdc-d5922c0c1db3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.581355] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9a3196c-8e75-483f-9957-b55c20c90222 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.610368] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9fd1fa0-01c2-4806-91da-b483c351916a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.619213] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b20805dd-a30b-49f3-87e5-f1a9d3a059e6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.633412] env[63531]: DEBUG nova.compute.provider_tree [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 989.740984] env[63531]: INFO nova.compute.manager [-] [instance: 0fb3b687-62cc-4296-b4ed-537aba18c880] Took 1.42 seconds to deallocate network for instance. [ 989.741771] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b8aa7e30-167c-41a5-934a-ef76c3a9cbd3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "3bf08a09-b6e6-4a58-8d35-7061a8e59afd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.166s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.776470] env[63531]: DEBUG oslo_vmware.api [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118358, 'name': ReconfigVM_Task, 'duration_secs': 0.51024} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.780057] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Reconfigured VM instance instance-0000005a to attach disk [datastore2] 4757625a-8398-4b2d-b6fe-40eea8913068/4757625a-8398-4b2d-b6fe-40eea8913068.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 989.780905] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e1e9e81c-72c3-4814-aca0-33f8f8bb94dc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.787604] env[63531]: DEBUG nova.network.neutron [-] [instance: baf2db4e-2cdb-47f5-9a5c-97a233a459c9] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.789124] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118359, 'name': CreateVM_Task} progress is 25%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.790731] env[63531]: DEBUG oslo_vmware.api [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for the task: (returnval){ [ 989.790731] env[63531]: value = "task-1118361" [ 989.790731] env[63531]: _type = "Task" [ 989.790731] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.801119] env[63531]: DEBUG oslo_vmware.api [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118361, 'name': Rename_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.846375] env[63531]: DEBUG oslo_vmware.api [None req-1b6ffe13-6d20-4c93-b882-81d0eee2d420 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52ff628f-1382-16ab-2c3d-71a8a65c35c7, 'name': SearchDatastore_Task, 'duration_secs': 0.008002} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.846690] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1b6ffe13-6d20-4c93-b882-81d0eee2d420 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.975761] env[63531]: DEBUG oslo_vmware.api [None req-ac717b99-ffee-40ff-8c50-9c164bf5d8ae tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118360, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.101462] env[63531]: DEBUG nova.network.neutron [req-ec79c6d9-99c8-42df-9dd8-2f9063496433 req-b0460b96-b2e0-4388-8231-81cdb09adf8a service nova] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Updated VIF entry in instance network info cache for port a926b26b-b99b-49f2-b7dd-6ca158ead545. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 990.101870] env[63531]: DEBUG nova.network.neutron [req-ec79c6d9-99c8-42df-9dd8-2f9063496433 req-b0460b96-b2e0-4388-8231-81cdb09adf8a service nova] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Updating instance_info_cache with network_info: [{"id": "a926b26b-b99b-49f2-b7dd-6ca158ead545", "address": "fa:16:3e:99:7c:93", "network": {"id": "caf86428-59a5-4e50-9135-2ab8473af64d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1673994696-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ba8b284854242c392aec5326e996239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa926b26b-b9", "ovs_interfaceid": "a926b26b-b99b-49f2-b7dd-6ca158ead545", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.136852] env[63531]: DEBUG nova.scheduler.client.report [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 990.250923] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9b48ef40-e853-4488-bc73-2ca4984fb08c tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.286660] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118359, 'name': CreateVM_Task, 'duration_secs': 0.757728} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.286981] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 990.288049] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.288353] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.288785] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 990.289162] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f1527e48-f54d-438e-a222-6e7850ff0131 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.291534] env[63531]: INFO nova.compute.manager [-] [instance: baf2db4e-2cdb-47f5-9a5c-97a233a459c9] Took 1.40 seconds to deallocate network for instance. [ 990.301268] env[63531]: DEBUG oslo_vmware.api [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for the task: (returnval){ [ 990.301268] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5265ee47-a88a-0c68-0f2a-36a50e4217d8" [ 990.301268] env[63531]: _type = "Task" [ 990.301268] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.308064] env[63531]: DEBUG oslo_vmware.api [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118361, 'name': Rename_Task, 'duration_secs': 0.191861} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.309094] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 990.309495] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a5281c02-461e-4606-b5da-59a1181e0c02 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.314394] env[63531]: DEBUG oslo_vmware.api [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5265ee47-a88a-0c68-0f2a-36a50e4217d8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.320023] env[63531]: DEBUG oslo_vmware.api [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for the task: (returnval){ [ 990.320023] env[63531]: value = "task-1118362" [ 990.320023] env[63531]: _type = "Task" [ 990.320023] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.328465] env[63531]: DEBUG oslo_vmware.api [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118362, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.476217] env[63531]: DEBUG oslo_vmware.api [None req-ac717b99-ffee-40ff-8c50-9c164bf5d8ae tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118360, 'name': PowerOnVM_Task, 'duration_secs': 0.612848} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.476217] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac717b99-ffee-40ff-8c50-9c164bf5d8ae tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 990.478409] env[63531]: DEBUG nova.compute.manager [None req-ac717b99-ffee-40ff-8c50-9c164bf5d8ae tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 990.479440] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc1298d6-2de4-4ea9-8ae5-cb6bd8831921 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.604690] env[63531]: DEBUG oslo_concurrency.lockutils [req-ec79c6d9-99c8-42df-9dd8-2f9063496433 req-b0460b96-b2e0-4388-8231-81cdb09adf8a service nova] Releasing lock "refresh_cache-e742d3af-5a45-4274-adfb-c6138bf84d6a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.605022] env[63531]: DEBUG nova.compute.manager [req-ec79c6d9-99c8-42df-9dd8-2f9063496433 req-b0460b96-b2e0-4388-8231-81cdb09adf8a service nova] [instance: 0fb3b687-62cc-4296-b4ed-537aba18c880] Received event network-vif-deleted-ac505e2f-66a8-467d-b109-d86556eedc79 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 990.605242] env[63531]: INFO nova.compute.manager [req-ec79c6d9-99c8-42df-9dd8-2f9063496433 req-b0460b96-b2e0-4388-8231-81cdb09adf8a service nova] [instance: 0fb3b687-62cc-4296-b4ed-537aba18c880] Neutron deleted interface ac505e2f-66a8-467d-b109-d86556eedc79; detaching it from the instance and deleting it from the info cache [ 990.605430] env[63531]: DEBUG nova.network.neutron [req-ec79c6d9-99c8-42df-9dd8-2f9063496433 req-b0460b96-b2e0-4388-8231-81cdb09adf8a service nova] [instance: 0fb3b687-62cc-4296-b4ed-537aba18c880] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.644737] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.525s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.645342] env[63531]: DEBUG nova.compute.manager [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 990.648243] env[63531]: DEBUG oslo_concurrency.lockutils [None req-2df5001e-3daa-4d21-82f3-a5fde08b46fb tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.510s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.648499] env[63531]: DEBUG nova.objects.instance [None req-2df5001e-3daa-4d21-82f3-a5fde08b46fb tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Lazy-loading 'resources' on Instance uuid 1ca7b851-2a06-4181-8271-58aafcd322d6 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 990.804107] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0f3b984b-d279-4944-8a9e-cb9bdebbf475 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.815455] env[63531]: DEBUG oslo_vmware.api [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5265ee47-a88a-0c68-0f2a-36a50e4217d8, 'name': SearchDatastore_Task, 'duration_secs': 0.020127} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.815766] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.816052] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 990.816443] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.816604] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.816792] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 990.817067] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3a5c2157-00cd-4d39-a157-dcbfeb5a8153 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.826757] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 990.826999] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 990.828159] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b96580c8-fa24-4ff9-8f8c-af959906e4b2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.834563] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4818e295-516e-436b-8abc-55b2e5232248 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "3bf08a09-b6e6-4a58-8d35-7061a8e59afd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.834849] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4818e295-516e-436b-8abc-55b2e5232248 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "3bf08a09-b6e6-4a58-8d35-7061a8e59afd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.835079] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4818e295-516e-436b-8abc-55b2e5232248 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "3bf08a09-b6e6-4a58-8d35-7061a8e59afd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.835273] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4818e295-516e-436b-8abc-55b2e5232248 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "3bf08a09-b6e6-4a58-8d35-7061a8e59afd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.836224] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4818e295-516e-436b-8abc-55b2e5232248 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "3bf08a09-b6e6-4a58-8d35-7061a8e59afd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.837188] env[63531]: DEBUG oslo_vmware.api [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118362, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.837650] env[63531]: INFO nova.compute.manager [None req-4818e295-516e-436b-8abc-55b2e5232248 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3bf08a09-b6e6-4a58-8d35-7061a8e59afd] Terminating instance [ 990.840760] env[63531]: DEBUG nova.compute.manager [None req-4818e295-516e-436b-8abc-55b2e5232248 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3bf08a09-b6e6-4a58-8d35-7061a8e59afd] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 990.840964] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-4818e295-516e-436b-8abc-55b2e5232248 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3bf08a09-b6e6-4a58-8d35-7061a8e59afd] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 990.841735] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99df7130-c10c-4fda-bd78-3b63b17cd2cb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.845540] env[63531]: DEBUG oslo_vmware.api [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for the task: (returnval){ [ 990.845540] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]521157de-74f9-ebea-7cea-454293ca326a" [ 990.845540] env[63531]: _type = "Task" [ 990.845540] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.850924] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-4818e295-516e-436b-8abc-55b2e5232248 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3bf08a09-b6e6-4a58-8d35-7061a8e59afd] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 990.851518] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-13e1e631-cb32-4256-8d0e-d1bdea59d630 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.856124] env[63531]: DEBUG oslo_vmware.api [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]521157de-74f9-ebea-7cea-454293ca326a, 'name': SearchDatastore_Task, 'duration_secs': 0.010338} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.857281] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5af19c28-9232-44b3-9a8a-422454b1e192 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.861024] env[63531]: DEBUG oslo_vmware.api [None req-4818e295-516e-436b-8abc-55b2e5232248 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 990.861024] env[63531]: value = "task-1118363" [ 990.861024] env[63531]: _type = "Task" [ 990.861024] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.862276] env[63531]: DEBUG oslo_vmware.api [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for the task: (returnval){ [ 990.862276] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52130539-7435-e5df-fe02-fff3e448f47f" [ 990.862276] env[63531]: _type = "Task" [ 990.862276] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.873031] env[63531]: DEBUG oslo_vmware.api [None req-4818e295-516e-436b-8abc-55b2e5232248 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118363, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.875954] env[63531]: DEBUG oslo_vmware.api [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52130539-7435-e5df-fe02-fff3e448f47f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.965595] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6b6dc263-18a2-4a55-883c-76c6818d2fb1 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "27c23b7c-a68f-43c9-a125-bd657feb3c5b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.965821] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6b6dc263-18a2-4a55-883c-76c6818d2fb1 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "27c23b7c-a68f-43c9-a125-bd657feb3c5b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.966067] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6b6dc263-18a2-4a55-883c-76c6818d2fb1 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "27c23b7c-a68f-43c9-a125-bd657feb3c5b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.966413] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6b6dc263-18a2-4a55-883c-76c6818d2fb1 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "27c23b7c-a68f-43c9-a125-bd657feb3c5b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.966620] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6b6dc263-18a2-4a55-883c-76c6818d2fb1 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "27c23b7c-a68f-43c9-a125-bd657feb3c5b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.970067] env[63531]: INFO nova.compute.manager [None req-6b6dc263-18a2-4a55-883c-76c6818d2fb1 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Terminating instance [ 990.972065] env[63531]: DEBUG nova.compute.manager [None req-6b6dc263-18a2-4a55-883c-76c6818d2fb1 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 990.972431] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-6b6dc263-18a2-4a55-883c-76c6818d2fb1 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 990.973220] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c87dbb3-a9c6-40db-830e-6387e8f84bd9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.982678] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b6dc263-18a2-4a55-883c-76c6818d2fb1 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 990.985016] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a9fad54a-f8da-43c7-a8c1-a163c3f4fa90 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.999548] env[63531]: DEBUG oslo_vmware.api [None req-6b6dc263-18a2-4a55-883c-76c6818d2fb1 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 990.999548] env[63531]: value = "task-1118364" [ 990.999548] env[63531]: _type = "Task" [ 990.999548] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.011063] env[63531]: DEBUG oslo_vmware.api [None req-6b6dc263-18a2-4a55-883c-76c6818d2fb1 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118364, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.108372] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1a52959a-eaea-4b89-bd31-0619113c139c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.117880] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ccd580e-eb50-4cf4-91da-20ca0eaf10d7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.152588] env[63531]: DEBUG nova.compute.utils [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 991.157018] env[63531]: DEBUG nova.compute.manager [req-ec79c6d9-99c8-42df-9dd8-2f9063496433 req-b0460b96-b2e0-4388-8231-81cdb09adf8a service nova] [instance: 0fb3b687-62cc-4296-b4ed-537aba18c880] Detach interface failed, port_id=ac505e2f-66a8-467d-b109-d86556eedc79, reason: Instance 0fb3b687-62cc-4296-b4ed-537aba18c880 could not be found. {{(pid=63531) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 991.158788] env[63531]: DEBUG nova.compute.manager [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 991.158989] env[63531]: DEBUG nova.network.neutron [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 991.229259] env[63531]: DEBUG nova.policy [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2ec68c1860c5484cb8094370baa80c98', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '57e9bf4ee0f1463fa67b0f8deff30fdc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 991.257046] env[63531]: DEBUG nova.compute.manager [req-af8be727-f5b5-4154-89cc-efc33d378103 req-149cab4a-60f5-4497-abb7-e38b221b179e service nova] [instance: baf2db4e-2cdb-47f5-9a5c-97a233a459c9] Received event network-vif-deleted-df2cde2b-a781-49ba-9120-bc742e62367d {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 991.333993] env[63531]: DEBUG oslo_vmware.api [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118362, 'name': PowerOnVM_Task, 'duration_secs': 0.829315} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.336426] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 991.337017] env[63531]: INFO nova.compute.manager [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Took 9.24 seconds to spawn the instance on the hypervisor. [ 991.338329] env[63531]: DEBUG nova.compute.manager [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 991.338423] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26ef8c4a-3a60-4a72-a671-d4b687dd72d8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.381428] env[63531]: DEBUG oslo_vmware.api [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52130539-7435-e5df-fe02-fff3e448f47f, 'name': SearchDatastore_Task, 'duration_secs': 0.012171} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.381428] env[63531]: DEBUG oslo_vmware.api [None req-4818e295-516e-436b-8abc-55b2e5232248 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118363, 'name': PowerOffVM_Task, 'duration_secs': 0.191081} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.381428] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.381428] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] e742d3af-5a45-4274-adfb-c6138bf84d6a/e742d3af-5a45-4274-adfb-c6138bf84d6a.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 991.381428] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-4818e295-516e-436b-8abc-55b2e5232248 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3bf08a09-b6e6-4a58-8d35-7061a8e59afd] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 991.381663] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-4818e295-516e-436b-8abc-55b2e5232248 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3bf08a09-b6e6-4a58-8d35-7061a8e59afd] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 991.381951] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7605767b-ac92-4bf8-b5a4-80d773a72f23 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.384146] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b29d5894-fd8f-40dc-8247-02b60bc63e88 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.392889] env[63531]: DEBUG oslo_vmware.api [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for the task: (returnval){ [ 991.392889] env[63531]: value = "task-1118365" [ 991.392889] env[63531]: _type = "Task" [ 991.392889] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.405408] env[63531]: DEBUG oslo_vmware.api [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118365, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.439506] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2773b47-346e-4c48-9605-9050d06fea36 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.449654] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b9b78f1-8064-485f-bcef-81b11998d22d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.454199] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-4818e295-516e-436b-8abc-55b2e5232248 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3bf08a09-b6e6-4a58-8d35-7061a8e59afd] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 991.454424] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-4818e295-516e-436b-8abc-55b2e5232248 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3bf08a09-b6e6-4a58-8d35-7061a8e59afd] Deleting contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 991.454609] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-4818e295-516e-436b-8abc-55b2e5232248 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Deleting the datastore file [datastore2] 3bf08a09-b6e6-4a58-8d35-7061a8e59afd {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 991.454848] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-48123d47-c3cb-4829-90a6-5711e40a8fcc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.483332] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e18b16aa-ac9a-4d35-baf8-e848d6e57054 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.487108] env[63531]: DEBUG oslo_vmware.api [None req-4818e295-516e-436b-8abc-55b2e5232248 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 991.487108] env[63531]: value = "task-1118367" [ 991.487108] env[63531]: _type = "Task" [ 991.487108] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.493843] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e90ff01-ed19-4586-83c3-177c16f5692b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.501869] env[63531]: DEBUG oslo_vmware.api [None req-4818e295-516e-436b-8abc-55b2e5232248 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118367, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.507953] env[63531]: DEBUG oslo_concurrency.lockutils [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.508228] env[63531]: DEBUG oslo_concurrency.lockutils [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.517543] env[63531]: INFO nova.compute.manager [None req-32dcc163-a5f1-4b05-af53-551cacd6f588 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Unrescuing [ 991.517777] env[63531]: DEBUG oslo_concurrency.lockutils [None req-32dcc163-a5f1-4b05-af53-551cacd6f588 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquiring lock "refresh_cache-ccf00b09-29fe-4a76-a8af-97774acd77de" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.517976] env[63531]: DEBUG oslo_concurrency.lockutils [None req-32dcc163-a5f1-4b05-af53-551cacd6f588 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquired lock "refresh_cache-ccf00b09-29fe-4a76-a8af-97774acd77de" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.518190] env[63531]: DEBUG nova.network.neutron [None req-32dcc163-a5f1-4b05-af53-551cacd6f588 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 991.519538] env[63531]: DEBUG nova.compute.provider_tree [None req-2df5001e-3daa-4d21-82f3-a5fde08b46fb tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 991.525808] env[63531]: DEBUG oslo_vmware.api [None req-6b6dc263-18a2-4a55-883c-76c6818d2fb1 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118364, 'name': PowerOffVM_Task, 'duration_secs': 0.19178} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.526655] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b6dc263-18a2-4a55-883c-76c6818d2fb1 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 991.526855] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-6b6dc263-18a2-4a55-883c-76c6818d2fb1 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 991.527132] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a3041dd6-d4b9-4483-a560-ccafae32fd11 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.627730] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-6b6dc263-18a2-4a55-883c-76c6818d2fb1 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 991.628038] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-6b6dc263-18a2-4a55-883c-76c6818d2fb1 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 991.628301] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-6b6dc263-18a2-4a55-883c-76c6818d2fb1 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Deleting the datastore file [datastore1] 27c23b7c-a68f-43c9-a125-bd657feb3c5b {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 991.628614] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-92bccf1f-739a-4590-ad08-7251870fbd3e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.636383] env[63531]: DEBUG oslo_vmware.api [None req-6b6dc263-18a2-4a55-883c-76c6818d2fb1 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 991.636383] env[63531]: value = "task-1118369" [ 991.636383] env[63531]: _type = "Task" [ 991.636383] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.645679] env[63531]: DEBUG oslo_vmware.api [None req-6b6dc263-18a2-4a55-883c-76c6818d2fb1 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118369, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.659584] env[63531]: DEBUG nova.compute.manager [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 991.700669] env[63531]: DEBUG nova.network.neutron [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Successfully created port: af2485f6-5ce2-4a83-9d95-7bea85cebff2 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 991.866063] env[63531]: INFO nova.compute.manager [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Took 24.26 seconds to build instance. [ 991.900661] env[63531]: DEBUG oslo_vmware.api [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118365, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.472701} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.900918] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] e742d3af-5a45-4274-adfb-c6138bf84d6a/e742d3af-5a45-4274-adfb-c6138bf84d6a.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 991.901259] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 991.901520] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f90b76e3-72b0-4a1f-8817-ee1e186dd967 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.907477] env[63531]: DEBUG oslo_vmware.api [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for the task: (returnval){ [ 991.907477] env[63531]: value = "task-1118370" [ 991.907477] env[63531]: _type = "Task" [ 991.907477] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.925813] env[63531]: DEBUG oslo_vmware.api [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118370, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.002796] env[63531]: DEBUG oslo_vmware.api [None req-4818e295-516e-436b-8abc-55b2e5232248 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118367, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.226558} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.003355] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-4818e295-516e-436b-8abc-55b2e5232248 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 992.003642] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-4818e295-516e-436b-8abc-55b2e5232248 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3bf08a09-b6e6-4a58-8d35-7061a8e59afd] Deleted contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 992.003983] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-4818e295-516e-436b-8abc-55b2e5232248 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3bf08a09-b6e6-4a58-8d35-7061a8e59afd] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 992.004332] env[63531]: INFO nova.compute.manager [None req-4818e295-516e-436b-8abc-55b2e5232248 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 3bf08a09-b6e6-4a58-8d35-7061a8e59afd] Took 1.16 seconds to destroy the instance on the hypervisor. [ 992.004756] env[63531]: DEBUG oslo.service.loopingcall [None req-4818e295-516e-436b-8abc-55b2e5232248 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 992.005114] env[63531]: DEBUG nova.compute.manager [-] [instance: 3bf08a09-b6e6-4a58-8d35-7061a8e59afd] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 992.005295] env[63531]: DEBUG nova.network.neutron [-] [instance: 3bf08a09-b6e6-4a58-8d35-7061a8e59afd] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 992.024186] env[63531]: DEBUG nova.compute.manager [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 992.032400] env[63531]: DEBUG nova.scheduler.client.report [None req-2df5001e-3daa-4d21-82f3-a5fde08b46fb tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 992.146634] env[63531]: DEBUG oslo_vmware.api [None req-6b6dc263-18a2-4a55-883c-76c6818d2fb1 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118369, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.357343} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.146858] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-6b6dc263-18a2-4a55-883c-76c6818d2fb1 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 992.147055] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-6b6dc263-18a2-4a55-883c-76c6818d2fb1 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 992.147250] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-6b6dc263-18a2-4a55-883c-76c6818d2fb1 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 992.147429] env[63531]: INFO nova.compute.manager [None req-6b6dc263-18a2-4a55-883c-76c6818d2fb1 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Took 1.18 seconds to destroy the instance on the hypervisor. [ 992.147675] env[63531]: DEBUG oslo.service.loopingcall [None req-6b6dc263-18a2-4a55-883c-76c6818d2fb1 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 992.147887] env[63531]: DEBUG nova.compute.manager [-] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 992.147998] env[63531]: DEBUG nova.network.neutron [-] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 992.366905] env[63531]: DEBUG oslo_concurrency.lockutils [None req-32efe528-3ccb-48dc-a935-cc8a73377829 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "4757625a-8398-4b2d-b6fe-40eea8913068" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.773s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.421175] env[63531]: DEBUG oslo_vmware.api [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118370, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.098819} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.422498] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 992.423376] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3c9d7c9-3650-4657-9824-d99fc00c8c26 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.455189] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Reconfiguring VM instance instance-0000005b to attach disk [datastore1] e742d3af-5a45-4274-adfb-c6138bf84d6a/e742d3af-5a45-4274-adfb-c6138bf84d6a.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 992.455896] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-82c7aa70-8478-43cc-a9ba-4416008bcd21 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.489245] env[63531]: DEBUG oslo_vmware.api [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for the task: (returnval){ [ 992.489245] env[63531]: value = "task-1118371" [ 992.489245] env[63531]: _type = "Task" [ 992.489245] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.498151] env[63531]: DEBUG oslo_vmware.api [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118371, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.538033] env[63531]: DEBUG oslo_concurrency.lockutils [None req-2df5001e-3daa-4d21-82f3-a5fde08b46fb tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.889s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.542246] env[63531]: DEBUG oslo_concurrency.lockutils [None req-fc3f1fb2-bd92-4abe-bcb8-7277fece7b3a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.914s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.542469] env[63531]: DEBUG nova.objects.instance [None req-fc3f1fb2-bd92-4abe-bcb8-7277fece7b3a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lazy-loading 'resources' on Instance uuid 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 992.557782] env[63531]: DEBUG nova.network.neutron [None req-32dcc163-a5f1-4b05-af53-551cacd6f588 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Updating instance_info_cache with network_info: [{"id": "7433ce46-ff5a-4e78-bcc9-9dcf2264a01b", "address": "fa:16:3e:ad:77:b5", "network": {"id": "913045d9-e90b-4366-987c-7706b73a6147", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1696510389-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f71c7db0a3614995becc25d9a8dce787", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c297fe21-cd0b-4226-813b-a65d2358d034", "external-id": "nsx-vlan-transportzone-98", "segmentation_id": 98, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7433ce46-ff", "ovs_interfaceid": "7433ce46-ff5a-4e78-bcc9-9dcf2264a01b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.561195] env[63531]: DEBUG oslo_concurrency.lockutils [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.567504] env[63531]: INFO nova.scheduler.client.report [None req-2df5001e-3daa-4d21-82f3-a5fde08b46fb tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Deleted allocations for instance 1ca7b851-2a06-4181-8271-58aafcd322d6 [ 992.628201] env[63531]: DEBUG nova.compute.manager [req-ff265d7f-e66f-47f2-ae39-8f741187dddf req-6d261302-f280-454d-923a-d93cd5a25349 service nova] [instance: 3bf08a09-b6e6-4a58-8d35-7061a8e59afd] Received event network-vif-deleted-aa2bee40-6eaf-4fa4-a313-0cf4ed60310d {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 992.628512] env[63531]: INFO nova.compute.manager [req-ff265d7f-e66f-47f2-ae39-8f741187dddf req-6d261302-f280-454d-923a-d93cd5a25349 service nova] [instance: 3bf08a09-b6e6-4a58-8d35-7061a8e59afd] Neutron deleted interface aa2bee40-6eaf-4fa4-a313-0cf4ed60310d; detaching it from the instance and deleting it from the info cache [ 992.628759] env[63531]: DEBUG nova.network.neutron [req-ff265d7f-e66f-47f2-ae39-8f741187dddf req-6d261302-f280-454d-923a-d93cd5a25349 service nova] [instance: 3bf08a09-b6e6-4a58-8d35-7061a8e59afd] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.671550] env[63531]: DEBUG nova.compute.manager [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 992.702110] env[63531]: DEBUG nova.virt.hardware [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 992.702459] env[63531]: DEBUG nova.virt.hardware [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 992.702631] env[63531]: DEBUG nova.virt.hardware [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 992.702825] env[63531]: DEBUG nova.virt.hardware [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 992.702981] env[63531]: DEBUG nova.virt.hardware [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 992.703167] env[63531]: DEBUG nova.virt.hardware [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 992.703466] env[63531]: DEBUG nova.virt.hardware [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 992.703528] env[63531]: DEBUG nova.virt.hardware [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 992.703681] env[63531]: DEBUG nova.virt.hardware [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 992.703851] env[63531]: DEBUG nova.virt.hardware [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 992.704041] env[63531]: DEBUG nova.virt.hardware [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 992.704921] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-940b48af-98e3-4896-8eaf-e9f4877c82b2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.714754] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4871104f-3b25-4452-bd18-836cbb111803 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.822535] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0f17d244-380a-4408-853a-1a2f399662e9 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquiring lock "911f23f7-e320-48f3-87a9-0239b013feff" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.822810] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0f17d244-380a-4408-853a-1a2f399662e9 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "911f23f7-e320-48f3-87a9-0239b013feff" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.915354] env[63531]: DEBUG nova.network.neutron [-] [instance: 3bf08a09-b6e6-4a58-8d35-7061a8e59afd] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.999586] env[63531]: DEBUG oslo_vmware.api [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118371, 'name': ReconfigVM_Task, 'duration_secs': 0.271443} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.999893] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Reconfigured VM instance instance-0000005b to attach disk [datastore1] e742d3af-5a45-4274-adfb-c6138bf84d6a/e742d3af-5a45-4274-adfb-c6138bf84d6a.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 993.000557] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6ef0b3a3-467b-43e3-a7c4-c6218ed4c7f2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.008965] env[63531]: DEBUG oslo_vmware.api [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for the task: (returnval){ [ 993.008965] env[63531]: value = "task-1118372" [ 993.008965] env[63531]: _type = "Task" [ 993.008965] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.013968] env[63531]: DEBUG nova.network.neutron [-] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 993.020982] env[63531]: DEBUG oslo_vmware.api [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118372, 'name': Rename_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.061205] env[63531]: DEBUG oslo_concurrency.lockutils [None req-32dcc163-a5f1-4b05-af53-551cacd6f588 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Releasing lock "refresh_cache-ccf00b09-29fe-4a76-a8af-97774acd77de" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.061925] env[63531]: DEBUG nova.objects.instance [None req-32dcc163-a5f1-4b05-af53-551cacd6f588 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Lazy-loading 'flavor' on Instance uuid ccf00b09-29fe-4a76-a8af-97774acd77de {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 993.081228] env[63531]: DEBUG oslo_concurrency.lockutils [None req-2df5001e-3daa-4d21-82f3-a5fde08b46fb tempest-ServersNegativeTestJSON-1199916099 tempest-ServersNegativeTestJSON-1199916099-project-member] Lock "1ca7b851-2a06-4181-8271-58aafcd322d6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.871s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.178774] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0ba55d37-e167-4fb2-b64e-05efcb1b1529 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.178774] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fdda927-58b8-4f89-9588-3992f0e0ab7f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.185356] env[63531]: DEBUG nova.compute.manager [req-ff265d7f-e66f-47f2-ae39-8f741187dddf req-6d261302-f280-454d-923a-d93cd5a25349 service nova] [instance: 3bf08a09-b6e6-4a58-8d35-7061a8e59afd] Detach interface failed, port_id=aa2bee40-6eaf-4fa4-a313-0cf4ed60310d, reason: Instance 3bf08a09-b6e6-4a58-8d35-7061a8e59afd could not be found. {{(pid=63531) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 993.309546] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1709b242-ee6f-4fec-90e3-70f9d4565557 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.318347] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78d7b31e-c36b-49e4-aa68-7d98018fc6b7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.325710] env[63531]: DEBUG nova.compute.utils [None req-0f17d244-380a-4408-853a-1a2f399662e9 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 993.353486] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0f17d244-380a-4408-853a-1a2f399662e9 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "911f23f7-e320-48f3-87a9-0239b013feff" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 0.531s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.354516] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0791a610-63ee-4d35-9bd7-02cada46c79f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.364020] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dfa9952-180c-4485-9245-7b584683ea08 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.380340] env[63531]: DEBUG nova.compute.provider_tree [None req-fc3f1fb2-bd92-4abe-bcb8-7277fece7b3a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 993.417767] env[63531]: INFO nova.compute.manager [-] [instance: 3bf08a09-b6e6-4a58-8d35-7061a8e59afd] Took 1.41 seconds to deallocate network for instance. [ 993.500959] env[63531]: DEBUG nova.network.neutron [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Successfully updated port: af2485f6-5ce2-4a83-9d95-7bea85cebff2 {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 993.516817] env[63531]: INFO nova.compute.manager [-] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Took 1.37 seconds to deallocate network for instance. [ 993.522426] env[63531]: DEBUG oslo_vmware.api [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118372, 'name': Rename_Task, 'duration_secs': 0.159256} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.527758] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 993.528257] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-13e55140-8563-40b2-a34b-d677202f98b8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.537945] env[63531]: DEBUG oslo_vmware.api [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for the task: (returnval){ [ 993.537945] env[63531]: value = "task-1118373" [ 993.537945] env[63531]: _type = "Task" [ 993.537945] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.555889] env[63531]: DEBUG oslo_vmware.api [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118373, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.568009] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c709773c-643d-4794-87a8-9cb42b93b89a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.573978] env[63531]: DEBUG nova.compute.manager [req-a032f026-4e59-4ea0-941d-da7988cc346f req-ce37b926-7dea-4c01-a98f-13cd3d591bb6 service nova] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Received event network-vif-plugged-af2485f6-5ce2-4a83-9d95-7bea85cebff2 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 993.574239] env[63531]: DEBUG oslo_concurrency.lockutils [req-a032f026-4e59-4ea0-941d-da7988cc346f req-ce37b926-7dea-4c01-a98f-13cd3d591bb6 service nova] Acquiring lock "18057005-1ff1-4e3d-bccc-8c89755d6db7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.574450] env[63531]: DEBUG oslo_concurrency.lockutils [req-a032f026-4e59-4ea0-941d-da7988cc346f req-ce37b926-7dea-4c01-a98f-13cd3d591bb6 service nova] Lock "18057005-1ff1-4e3d-bccc-8c89755d6db7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.574619] env[63531]: DEBUG oslo_concurrency.lockutils [req-a032f026-4e59-4ea0-941d-da7988cc346f req-ce37b926-7dea-4c01-a98f-13cd3d591bb6 service nova] Lock "18057005-1ff1-4e3d-bccc-8c89755d6db7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.574792] env[63531]: DEBUG nova.compute.manager [req-a032f026-4e59-4ea0-941d-da7988cc346f req-ce37b926-7dea-4c01-a98f-13cd3d591bb6 service nova] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] No waiting events found dispatching network-vif-plugged-af2485f6-5ce2-4a83-9d95-7bea85cebff2 {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 993.574960] env[63531]: WARNING nova.compute.manager [req-a032f026-4e59-4ea0-941d-da7988cc346f req-ce37b926-7dea-4c01-a98f-13cd3d591bb6 service nova] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Received unexpected event network-vif-plugged-af2485f6-5ce2-4a83-9d95-7bea85cebff2 for instance with vm_state building and task_state spawning. [ 993.603904] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-32dcc163-a5f1-4b05-af53-551cacd6f588 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 993.604286] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cbae20f5-7b8b-4ad4-abbc-34ad21b67ee9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.612897] env[63531]: DEBUG oslo_vmware.api [None req-32dcc163-a5f1-4b05-af53-551cacd6f588 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 993.612897] env[63531]: value = "task-1118374" [ 993.612897] env[63531]: _type = "Task" [ 993.612897] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.622991] env[63531]: DEBUG oslo_vmware.api [None req-32dcc163-a5f1-4b05-af53-551cacd6f588 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118374, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.885895] env[63531]: DEBUG nova.scheduler.client.report [None req-fc3f1fb2-bd92-4abe-bcb8-7277fece7b3a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 993.924283] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4818e295-516e-436b-8abc-55b2e5232248 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.976156] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquiring lock "refresh_cache-18057005-1ff1-4e3d-bccc-8c89755d6db7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.976365] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquired lock "refresh_cache-18057005-1ff1-4e3d-bccc-8c89755d6db7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.976556] env[63531]: DEBUG nova.network.neutron [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 994.029817] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6b6dc263-18a2-4a55-883c-76c6818d2fb1 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.051398] env[63531]: DEBUG oslo_vmware.api [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118373, 'name': PowerOnVM_Task, 'duration_secs': 0.484955} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.051813] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 994.052156] env[63531]: INFO nova.compute.manager [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Took 8.16 seconds to spawn the instance on the hypervisor. [ 994.052466] env[63531]: DEBUG nova.compute.manager [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 994.054049] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49918f2e-d69b-4903-b5cd-a4eaa6e50fd4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.129855] env[63531]: DEBUG oslo_vmware.api [None req-32dcc163-a5f1-4b05-af53-551cacd6f588 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118374, 'name': PowerOffVM_Task, 'duration_secs': 0.359508} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.130318] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-32dcc163-a5f1-4b05-af53-551cacd6f588 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 994.139241] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-32dcc163-a5f1-4b05-af53-551cacd6f588 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Reconfiguring VM instance instance-00000051 to detach disk 2002 {{(pid=63531) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 994.139654] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-76ecdf23-1a43-48a6-817c-17f41792cd50 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.170021] env[63531]: DEBUG oslo_vmware.api [None req-32dcc163-a5f1-4b05-af53-551cacd6f588 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 994.170021] env[63531]: value = "task-1118375" [ 994.170021] env[63531]: _type = "Task" [ 994.170021] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.182424] env[63531]: DEBUG oslo_vmware.api [None req-32dcc163-a5f1-4b05-af53-551cacd6f588 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118375, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.392341] env[63531]: DEBUG oslo_concurrency.lockutils [None req-fc3f1fb2-bd92-4abe-bcb8-7277fece7b3a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.849s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.393664] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.890s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.395265] env[63531]: INFO nova.compute.claims [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 994.405105] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0f17d244-380a-4408-853a-1a2f399662e9 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquiring lock "911f23f7-e320-48f3-87a9-0239b013feff" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.405363] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0f17d244-380a-4408-853a-1a2f399662e9 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "911f23f7-e320-48f3-87a9-0239b013feff" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.405601] env[63531]: INFO nova.compute.manager [None req-0f17d244-380a-4408-853a-1a2f399662e9 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Attaching volume 3ac35683-e14e-4d02-8b0f-a5ea34354c7f to /dev/sdb [ 994.420476] env[63531]: INFO nova.scheduler.client.report [None req-fc3f1fb2-bd92-4abe-bcb8-7277fece7b3a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Deleted allocations for instance 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe [ 994.447590] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c23ce83-5ef3-4a67-9268-9a96ac065819 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.456769] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d25c94df-44ee-418f-a81a-d849d5f85792 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.471887] env[63531]: DEBUG nova.virt.block_device [None req-0f17d244-380a-4408-853a-1a2f399662e9 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Updating existing volume attachment record: 19a5fba7-6b26-4016-afd8-3b525f99e711 {{(pid=63531) _volume_attach /opt/stack/nova/nova/virt/block_device.py:679}} [ 994.536085] env[63531]: DEBUG nova.network.neutron [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 994.577860] env[63531]: INFO nova.compute.manager [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Took 22.37 seconds to build instance. [ 994.679106] env[63531]: DEBUG oslo_vmware.api [None req-32dcc163-a5f1-4b05-af53-551cacd6f588 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118375, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.772164] env[63531]: DEBUG nova.compute.manager [req-41531b4a-bc1a-4a1c-9f82-3b245576657e req-c47f6066-bcc1-4d89-ac8a-a71f492d7989 service nova] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Received event network-vif-deleted-a755a88f-c86c-4183-b9fa-a6133060ea5e {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 994.772164] env[63531]: DEBUG nova.compute.manager [req-41531b4a-bc1a-4a1c-9f82-3b245576657e req-c47f6066-bcc1-4d89-ac8a-a71f492d7989 service nova] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Received event network-changed-b37a092f-e03c-417f-bac1-3b2ebb5253d1 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 994.772164] env[63531]: DEBUG nova.compute.manager [req-41531b4a-bc1a-4a1c-9f82-3b245576657e req-c47f6066-bcc1-4d89-ac8a-a71f492d7989 service nova] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Refreshing instance network info cache due to event network-changed-b37a092f-e03c-417f-bac1-3b2ebb5253d1. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 994.772164] env[63531]: DEBUG oslo_concurrency.lockutils [req-41531b4a-bc1a-4a1c-9f82-3b245576657e req-c47f6066-bcc1-4d89-ac8a-a71f492d7989 service nova] Acquiring lock "refresh_cache-4757625a-8398-4b2d-b6fe-40eea8913068" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 994.772164] env[63531]: DEBUG oslo_concurrency.lockutils [req-41531b4a-bc1a-4a1c-9f82-3b245576657e req-c47f6066-bcc1-4d89-ac8a-a71f492d7989 service nova] Acquired lock "refresh_cache-4757625a-8398-4b2d-b6fe-40eea8913068" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 994.772164] env[63531]: DEBUG nova.network.neutron [req-41531b4a-bc1a-4a1c-9f82-3b245576657e req-c47f6066-bcc1-4d89-ac8a-a71f492d7989 service nova] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Refreshing network info cache for port b37a092f-e03c-417f-bac1-3b2ebb5253d1 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 994.903493] env[63531]: DEBUG nova.network.neutron [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Updating instance_info_cache with network_info: [{"id": "af2485f6-5ce2-4a83-9d95-7bea85cebff2", "address": "fa:16:3e:f0:7d:92", "network": {"id": "c4f4c994-36ab-42ab-94f8-78f29cbafa50", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-994507698-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "57e9bf4ee0f1463fa67b0f8deff30fdc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "767a3a48-41d4-4a0c-961d-0024837f63bd", "external-id": "nsx-vlan-transportzone-141", "segmentation_id": 141, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaf2485f6-5c", "ovs_interfaceid": "af2485f6-5ce2-4a83-9d95-7bea85cebff2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 994.933707] env[63531]: DEBUG oslo_concurrency.lockutils [None req-fc3f1fb2-bd92-4abe-bcb8-7277fece7b3a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.405s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.080565] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62f8efee-f802-4c1c-8891-2c8c006560a6 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "e742d3af-5a45-4274-adfb-c6138bf84d6a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.197s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.181868] env[63531]: DEBUG oslo_vmware.api [None req-32dcc163-a5f1-4b05-af53-551cacd6f588 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118375, 'name': ReconfigVM_Task, 'duration_secs': 0.778322} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.182357] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-32dcc163-a5f1-4b05-af53-551cacd6f588 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Reconfigured VM instance instance-00000051 to detach disk 2002 {{(pid=63531) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 995.184736] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-32dcc163-a5f1-4b05-af53-551cacd6f588 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 995.184736] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a5ee5376-841a-4b6d-9e94-838a61de0807 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.192114] env[63531]: DEBUG oslo_vmware.api [None req-32dcc163-a5f1-4b05-af53-551cacd6f588 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 995.192114] env[63531]: value = "task-1118379" [ 995.192114] env[63531]: _type = "Task" [ 995.192114] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.201982] env[63531]: DEBUG oslo_vmware.api [None req-32dcc163-a5f1-4b05-af53-551cacd6f588 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118379, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.408618] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Releasing lock "refresh_cache-18057005-1ff1-4e3d-bccc-8c89755d6db7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 995.408618] env[63531]: DEBUG nova.compute.manager [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Instance network_info: |[{"id": "af2485f6-5ce2-4a83-9d95-7bea85cebff2", "address": "fa:16:3e:f0:7d:92", "network": {"id": "c4f4c994-36ab-42ab-94f8-78f29cbafa50", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-994507698-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "57e9bf4ee0f1463fa67b0f8deff30fdc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "767a3a48-41d4-4a0c-961d-0024837f63bd", "external-id": "nsx-vlan-transportzone-141", "segmentation_id": 141, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaf2485f6-5c", "ovs_interfaceid": "af2485f6-5ce2-4a83-9d95-7bea85cebff2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 995.410798] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f0:7d:92', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '767a3a48-41d4-4a0c-961d-0024837f63bd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'af2485f6-5ce2-4a83-9d95-7bea85cebff2', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 995.419609] env[63531]: DEBUG oslo.service.loopingcall [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 995.420597] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 995.420969] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4dab2a55-078d-41e8-a3dd-18a9dc528a07 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.447707] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 995.447707] env[63531]: value = "task-1118380" [ 995.447707] env[63531]: _type = "Task" [ 995.447707] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.462092] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118380, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.632211] env[63531]: DEBUG nova.compute.manager [req-945badd7-2770-4563-9323-3aa8011207e4 req-5519317a-9578-43a4-9644-e3e2c496095b service nova] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Received event network-changed-af2485f6-5ce2-4a83-9d95-7bea85cebff2 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 995.632596] env[63531]: DEBUG nova.compute.manager [req-945badd7-2770-4563-9323-3aa8011207e4 req-5519317a-9578-43a4-9644-e3e2c496095b service nova] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Refreshing instance network info cache due to event network-changed-af2485f6-5ce2-4a83-9d95-7bea85cebff2. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 995.633132] env[63531]: DEBUG oslo_concurrency.lockutils [req-945badd7-2770-4563-9323-3aa8011207e4 req-5519317a-9578-43a4-9644-e3e2c496095b service nova] Acquiring lock "refresh_cache-18057005-1ff1-4e3d-bccc-8c89755d6db7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.633421] env[63531]: DEBUG oslo_concurrency.lockutils [req-945badd7-2770-4563-9323-3aa8011207e4 req-5519317a-9578-43a4-9644-e3e2c496095b service nova] Acquired lock "refresh_cache-18057005-1ff1-4e3d-bccc-8c89755d6db7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.633727] env[63531]: DEBUG nova.network.neutron [req-945badd7-2770-4563-9323-3aa8011207e4 req-5519317a-9578-43a4-9644-e3e2c496095b service nova] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Refreshing network info cache for port af2485f6-5ce2-4a83-9d95-7bea85cebff2 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 995.707625] env[63531]: DEBUG oslo_vmware.api [None req-32dcc163-a5f1-4b05-af53-551cacd6f588 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118379, 'name': PowerOnVM_Task, 'duration_secs': 0.453647} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.710752] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-32dcc163-a5f1-4b05-af53-551cacd6f588 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 995.713078] env[63531]: DEBUG nova.compute.manager [None req-32dcc163-a5f1-4b05-af53-551cacd6f588 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 995.714553] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e569d05-5f47-4f07-9f8b-bc216b7aadb8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.739612] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5794b6cf-6c1f-4d1b-ac16-9d8d024a8ed0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.748448] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b6a20c8-5d36-4ea7-86f5-2799b739c69c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.785757] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a933c3e2-caab-4c98-8f43-29ae8b189832 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.795044] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abc383da-61c1-4772-91d4-87836a15ca31 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.813590] env[63531]: DEBUG nova.compute.provider_tree [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 995.960435] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118380, 'name': CreateVM_Task, 'duration_secs': 0.412515} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.960435] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 995.960948] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.961147] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.961483] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 995.961750] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4cf2e9cc-af57-4210-9a51-5ef7fde92a72 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.967510] env[63531]: DEBUG oslo_vmware.api [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 995.967510] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]526777a6-8e27-4c64-e2e8-a4334d3fe15a" [ 995.967510] env[63531]: _type = "Task" [ 995.967510] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.977174] env[63531]: DEBUG oslo_vmware.api [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]526777a6-8e27-4c64-e2e8-a4334d3fe15a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.007568] env[63531]: DEBUG nova.network.neutron [req-41531b4a-bc1a-4a1c-9f82-3b245576657e req-c47f6066-bcc1-4d89-ac8a-a71f492d7989 service nova] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Updated VIF entry in instance network info cache for port b37a092f-e03c-417f-bac1-3b2ebb5253d1. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 996.009159] env[63531]: DEBUG nova.network.neutron [req-41531b4a-bc1a-4a1c-9f82-3b245576657e req-c47f6066-bcc1-4d89-ac8a-a71f492d7989 service nova] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Updating instance_info_cache with network_info: [{"id": "b37a092f-e03c-417f-bac1-3b2ebb5253d1", "address": "fa:16:3e:1e:cf:08", "network": {"id": "95ed2c7d-3d1a-4736-8e90-104002f5c40b", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-73549344-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.248", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9056486e3b640d0bcdb9369ed5d67ad", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04e15990-16e1-4cb2-b0f0-06c362e68c5e", "external-id": "nsx-vlan-transportzone-555", "segmentation_id": 555, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb37a092f-e0", "ovs_interfaceid": "b37a092f-e03c-417f-bac1-3b2ebb5253d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.317506] env[63531]: DEBUG nova.scheduler.client.report [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 996.483580] env[63531]: DEBUG oslo_vmware.api [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]526777a6-8e27-4c64-e2e8-a4334d3fe15a, 'name': SearchDatastore_Task, 'duration_secs': 0.011567} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.485470] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.485889] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 996.486259] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 996.486538] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.486828] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 996.487218] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-56a59ff5-72c1-46e1-a365-7f03e84c593a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.499943] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 996.499943] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 996.499943] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c78b186b-d94f-408d-ac53-bbe722a10bdf {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.512418] env[63531]: DEBUG oslo_concurrency.lockutils [req-41531b4a-bc1a-4a1c-9f82-3b245576657e req-c47f6066-bcc1-4d89-ac8a-a71f492d7989 service nova] Releasing lock "refresh_cache-4757625a-8398-4b2d-b6fe-40eea8913068" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.514350] env[63531]: DEBUG oslo_vmware.api [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 996.514350] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52e5a200-22bf-2cd5-9bd7-96f4baf085d0" [ 996.514350] env[63531]: _type = "Task" [ 996.514350] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.526386] env[63531]: DEBUG oslo_vmware.api [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52e5a200-22bf-2cd5-9bd7-96f4baf085d0, 'name': SearchDatastore_Task} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.527493] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f056b616-a19d-47cd-9453-1119ba1c6407 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.533968] env[63531]: DEBUG oslo_vmware.api [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 996.533968] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]527e495e-937a-ade2-0162-4ec7dce5e86c" [ 996.533968] env[63531]: _type = "Task" [ 996.533968] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.544383] env[63531]: DEBUG oslo_vmware.api [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]527e495e-937a-ade2-0162-4ec7dce5e86c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.677542] env[63531]: DEBUG nova.network.neutron [req-945badd7-2770-4563-9323-3aa8011207e4 req-5519317a-9578-43a4-9644-e3e2c496095b service nova] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Updated VIF entry in instance network info cache for port af2485f6-5ce2-4a83-9d95-7bea85cebff2. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 996.677958] env[63531]: DEBUG nova.network.neutron [req-945badd7-2770-4563-9323-3aa8011207e4 req-5519317a-9578-43a4-9644-e3e2c496095b service nova] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Updating instance_info_cache with network_info: [{"id": "af2485f6-5ce2-4a83-9d95-7bea85cebff2", "address": "fa:16:3e:f0:7d:92", "network": {"id": "c4f4c994-36ab-42ab-94f8-78f29cbafa50", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-994507698-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "57e9bf4ee0f1463fa67b0f8deff30fdc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "767a3a48-41d4-4a0c-961d-0024837f63bd", "external-id": "nsx-vlan-transportzone-141", "segmentation_id": 141, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaf2485f6-5c", "ovs_interfaceid": "af2485f6-5ce2-4a83-9d95-7bea85cebff2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.828997] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.435s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.829568] env[63531]: DEBUG nova.compute.manager [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 996.832398] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1b6ffe13-6d20-4c93-b882-81d0eee2d420 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 6.986s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.047988] env[63531]: DEBUG oslo_vmware.api [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]527e495e-937a-ade2-0162-4ec7dce5e86c, 'name': SearchDatastore_Task, 'duration_secs': 0.010407} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.048543] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.049269] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 18057005-1ff1-4e3d-bccc-8c89755d6db7/18057005-1ff1-4e3d-bccc-8c89755d6db7.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 997.049571] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-90c2de2f-490f-4738-84f6-3727695e6b67 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.059875] env[63531]: DEBUG oslo_vmware.api [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 997.059875] env[63531]: value = "task-1118382" [ 997.059875] env[63531]: _type = "Task" [ 997.059875] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.074800] env[63531]: DEBUG oslo_vmware.api [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118382, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.181348] env[63531]: DEBUG oslo_concurrency.lockutils [req-945badd7-2770-4563-9323-3aa8011207e4 req-5519317a-9578-43a4-9644-e3e2c496095b service nova] Releasing lock "refresh_cache-18057005-1ff1-4e3d-bccc-8c89755d6db7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.338058] env[63531]: DEBUG nova.compute.utils [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 997.346170] env[63531]: DEBUG nova.compute.manager [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 997.346170] env[63531]: DEBUG nova.network.neutron [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 997.449497] env[63531]: DEBUG nova.policy [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ff7b1cf5f24e47cba315f4c4ee47e131', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ee5fd84e1da04fd984f81578e4085850', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 997.570765] env[63531]: DEBUG oslo_vmware.api [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118382, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.474235} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.573233] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 18057005-1ff1-4e3d-bccc-8c89755d6db7/18057005-1ff1-4e3d-bccc-8c89755d6db7.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 997.573497] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 997.573929] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-14ed8211-82ab-4e1f-8edb-65a920486560 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.581889] env[63531]: DEBUG oslo_vmware.api [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 997.581889] env[63531]: value = "task-1118383" [ 997.581889] env[63531]: _type = "Task" [ 997.581889] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.592835] env[63531]: DEBUG oslo_vmware.api [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118383, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.598855] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3aed6d4a-d156-440b-8d3e-63048e33a761 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.606961] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4994639-ad4e-495e-9238-a952064afdfb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.640685] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2faf4e4-3969-41ae-b536-efe68deafc82 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.650076] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55df0640-e350-46b4-aa05-f4e32f84e317 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.665508] env[63531]: DEBUG nova.compute.provider_tree [None req-1b6ffe13-6d20-4c93-b882-81d0eee2d420 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 997.743369] env[63531]: DEBUG nova.compute.manager [req-4b784b8f-47e5-43af-be21-5383b3c13607 req-a10fc64b-68ea-44ff-849c-c988bc597b6a service nova] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Received event network-changed-a926b26b-b99b-49f2-b7dd-6ca158ead545 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 997.743619] env[63531]: DEBUG nova.compute.manager [req-4b784b8f-47e5-43af-be21-5383b3c13607 req-a10fc64b-68ea-44ff-849c-c988bc597b6a service nova] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Refreshing instance network info cache due to event network-changed-a926b26b-b99b-49f2-b7dd-6ca158ead545. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 997.743898] env[63531]: DEBUG oslo_concurrency.lockutils [req-4b784b8f-47e5-43af-be21-5383b3c13607 req-a10fc64b-68ea-44ff-849c-c988bc597b6a service nova] Acquiring lock "refresh_cache-e742d3af-5a45-4274-adfb-c6138bf84d6a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 997.744088] env[63531]: DEBUG oslo_concurrency.lockutils [req-4b784b8f-47e5-43af-be21-5383b3c13607 req-a10fc64b-68ea-44ff-849c-c988bc597b6a service nova] Acquired lock "refresh_cache-e742d3af-5a45-4274-adfb-c6138bf84d6a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 997.744320] env[63531]: DEBUG nova.network.neutron [req-4b784b8f-47e5-43af-be21-5383b3c13607 req-a10fc64b-68ea-44ff-849c-c988bc597b6a service nova] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Refreshing network info cache for port a926b26b-b99b-49f2-b7dd-6ca158ead545 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 997.845665] env[63531]: DEBUG nova.compute.manager [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 998.013052] env[63531]: DEBUG nova.network.neutron [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Successfully created port: 40adf4b3-527a-4414-b2dc-a24a39b897c6 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 998.043683] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "3fe7a7a8-fc43-4560-b101-227907c91fd0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.044182] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "3fe7a7a8-fc43-4560-b101-227907c91fd0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.092598] env[63531]: DEBUG oslo_vmware.api [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118383, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.187068} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.093120] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 998.093853] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a0a7df7-43a6-4431-a758-5038414ee5b7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.119265] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Reconfiguring VM instance instance-0000005c to attach disk [datastore1] 18057005-1ff1-4e3d-bccc-8c89755d6db7/18057005-1ff1-4e3d-bccc-8c89755d6db7.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 998.119576] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eb5e1d5f-a13d-47a4-88f2-7b73a8937cb2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.142131] env[63531]: DEBUG oslo_vmware.api [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 998.142131] env[63531]: value = "task-1118384" [ 998.142131] env[63531]: _type = "Task" [ 998.142131] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.151844] env[63531]: DEBUG oslo_vmware.api [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118384, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.169208] env[63531]: DEBUG nova.scheduler.client.report [None req-1b6ffe13-6d20-4c93-b882-81d0eee2d420 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 998.529376] env[63531]: DEBUG nova.network.neutron [req-4b784b8f-47e5-43af-be21-5383b3c13607 req-a10fc64b-68ea-44ff-849c-c988bc597b6a service nova] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Updated VIF entry in instance network info cache for port a926b26b-b99b-49f2-b7dd-6ca158ead545. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 998.529636] env[63531]: DEBUG nova.network.neutron [req-4b784b8f-47e5-43af-be21-5383b3c13607 req-a10fc64b-68ea-44ff-849c-c988bc597b6a service nova] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Updating instance_info_cache with network_info: [{"id": "a926b26b-b99b-49f2-b7dd-6ca158ead545", "address": "fa:16:3e:99:7c:93", "network": {"id": "caf86428-59a5-4e50-9135-2ab8473af64d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1673994696-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.168", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ba8b284854242c392aec5326e996239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa926b26b-b9", "ovs_interfaceid": "a926b26b-b99b-49f2-b7dd-6ca158ead545", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 998.546748] env[63531]: DEBUG nova.compute.manager [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 998.656916] env[63531]: DEBUG oslo_vmware.api [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118384, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.855323] env[63531]: DEBUG nova.compute.manager [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 998.881567] env[63531]: DEBUG nova.virt.hardware [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 998.881821] env[63531]: DEBUG nova.virt.hardware [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 998.881986] env[63531]: DEBUG nova.virt.hardware [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 998.882193] env[63531]: DEBUG nova.virt.hardware [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 998.882351] env[63531]: DEBUG nova.virt.hardware [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 998.882508] env[63531]: DEBUG nova.virt.hardware [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 998.882728] env[63531]: DEBUG nova.virt.hardware [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 998.882895] env[63531]: DEBUG nova.virt.hardware [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 998.883085] env[63531]: DEBUG nova.virt.hardware [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 998.883262] env[63531]: DEBUG nova.virt.hardware [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 998.883445] env[63531]: DEBUG nova.virt.hardware [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 998.884343] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-989d9e00-36b2-4c8c-bac4-c4ed1cccc18c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.892776] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3e3399e-c40b-40db-bb82-af15e3a6f1be {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.032658] env[63531]: DEBUG oslo_concurrency.lockutils [req-4b784b8f-47e5-43af-be21-5383b3c13607 req-a10fc64b-68ea-44ff-849c-c988bc597b6a service nova] Releasing lock "refresh_cache-e742d3af-5a45-4274-adfb-c6138bf84d6a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 999.069036] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.154522] env[63531]: DEBUG oslo_vmware.api [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118384, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.180277] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1b6ffe13-6d20-4c93-b882-81d0eee2d420 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.348s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.183218] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9b48ef40-e853-4488-bc73-2ca4984fb08c tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.933s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.183467] env[63531]: DEBUG nova.objects.instance [None req-9b48ef40-e853-4488-bc73-2ca4984fb08c tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Lazy-loading 'resources' on Instance uuid 0fb3b687-62cc-4296-b4ed-537aba18c880 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 999.510261] env[63531]: DEBUG nova.network.neutron [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Successfully updated port: 40adf4b3-527a-4414-b2dc-a24a39b897c6 {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 999.521913] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-0f17d244-380a-4408-853a-1a2f399662e9 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Volume attach. Driver type: vmdk {{(pid=63531) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 999.521913] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-0f17d244-380a-4408-853a-1a2f399662e9 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244776', 'volume_id': '3ac35683-e14e-4d02-8b0f-a5ea34354c7f', 'name': 'volume-3ac35683-e14e-4d02-8b0f-a5ea34354c7f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '911f23f7-e320-48f3-87a9-0239b013feff', 'attached_at': '', 'detached_at': '', 'volume_id': '3ac35683-e14e-4d02-8b0f-a5ea34354c7f', 'serial': '3ac35683-e14e-4d02-8b0f-a5ea34354c7f'} {{(pid=63531) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 999.523573] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6aa70dd-9ed7-4306-8e20-4cd104048a57 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.544844] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7d41992-537f-456c-b8d6-8ea5edaaa122 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.570230] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-0f17d244-380a-4408-853a-1a2f399662e9 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Reconfiguring VM instance instance-00000055 to attach disk [datastore1] volume-3ac35683-e14e-4d02-8b0f-a5ea34354c7f/volume-3ac35683-e14e-4d02-8b0f-a5ea34354c7f.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 999.570548] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a7fdde50-2794-4b59-ad72-2ce2e4edbbae {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.591526] env[63531]: DEBUG oslo_vmware.api [None req-0f17d244-380a-4408-853a-1a2f399662e9 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 999.591526] env[63531]: value = "task-1118385" [ 999.591526] env[63531]: _type = "Task" [ 999.591526] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.602154] env[63531]: DEBUG oslo_vmware.api [None req-0f17d244-380a-4408-853a-1a2f399662e9 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118385, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.655043] env[63531]: DEBUG oslo_vmware.api [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118384, 'name': ReconfigVM_Task, 'duration_secs': 1.064931} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.655324] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Reconfigured VM instance instance-0000005c to attach disk [datastore1] 18057005-1ff1-4e3d-bccc-8c89755d6db7/18057005-1ff1-4e3d-bccc-8c89755d6db7.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 999.656131] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-125b3612-da9c-46ba-9389-5b65b8568074 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.663642] env[63531]: DEBUG oslo_vmware.api [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 999.663642] env[63531]: value = "task-1118386" [ 999.663642] env[63531]: _type = "Task" [ 999.663642] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.674871] env[63531]: DEBUG oslo_vmware.api [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118386, 'name': Rename_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.735413] env[63531]: INFO nova.scheduler.client.report [None req-1b6ffe13-6d20-4c93-b882-81d0eee2d420 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Deleted allocation for migration 731cf9af-2cef-4013-8292-6c5aedf6f9ee [ 999.870341] env[63531]: DEBUG nova.compute.manager [req-276d1c3b-47a9-44ce-964d-3f1040caea71 req-d296b6cb-0de3-4ba3-8e8e-c2782bba6176 service nova] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Received event network-changed-7433ce46-ff5a-4e78-bcc9-9dcf2264a01b {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 999.870562] env[63531]: DEBUG nova.compute.manager [req-276d1c3b-47a9-44ce-964d-3f1040caea71 req-d296b6cb-0de3-4ba3-8e8e-c2782bba6176 service nova] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Refreshing instance network info cache due to event network-changed-7433ce46-ff5a-4e78-bcc9-9dcf2264a01b. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 999.870766] env[63531]: DEBUG oslo_concurrency.lockutils [req-276d1c3b-47a9-44ce-964d-3f1040caea71 req-d296b6cb-0de3-4ba3-8e8e-c2782bba6176 service nova] Acquiring lock "refresh_cache-ccf00b09-29fe-4a76-a8af-97774acd77de" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 999.871113] env[63531]: DEBUG oslo_concurrency.lockutils [req-276d1c3b-47a9-44ce-964d-3f1040caea71 req-d296b6cb-0de3-4ba3-8e8e-c2782bba6176 service nova] Acquired lock "refresh_cache-ccf00b09-29fe-4a76-a8af-97774acd77de" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 999.871317] env[63531]: DEBUG nova.network.neutron [req-276d1c3b-47a9-44ce-964d-3f1040caea71 req-d296b6cb-0de3-4ba3-8e8e-c2782bba6176 service nova] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Refreshing network info cache for port 7433ce46-ff5a-4e78-bcc9-9dcf2264a01b {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 999.935769] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d32012af-52ce-41f4-a3c6-7a668a21d149 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.944926] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78758238-c8d3-4e28-afc4-d4dc00d13dc8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.980695] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb6af42b-cd27-4880-9c4a-d6afc628832b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.989576] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d292f44d-95b1-49dc-8505-0fcb868bf19a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.005214] env[63531]: DEBUG nova.compute.provider_tree [None req-9b48ef40-e853-4488-bc73-2ca4984fb08c tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1000.013456] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Acquiring lock "refresh_cache-b7b2b7ac-2321-4cf3-b2ed-4877b893af63" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.013788] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Acquired lock "refresh_cache-b7b2b7ac-2321-4cf3-b2ed-4877b893af63" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.013788] env[63531]: DEBUG nova.network.neutron [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1000.101920] env[63531]: DEBUG oslo_vmware.api [None req-0f17d244-380a-4408-853a-1a2f399662e9 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118385, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.174348] env[63531]: DEBUG oslo_vmware.api [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118386, 'name': Rename_Task, 'duration_secs': 0.271831} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.175029] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1000.175029] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6795c8cf-0760-45ab-84d0-8269bfe334ae {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.183334] env[63531]: DEBUG oslo_vmware.api [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 1000.183334] env[63531]: value = "task-1118387" [ 1000.183334] env[63531]: _type = "Task" [ 1000.183334] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.192594] env[63531]: DEBUG oslo_vmware.api [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118387, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.242261] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1b6ffe13-6d20-4c93-b882-81d0eee2d420 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "29dd6de3-2765-408c-acc0-da47e5e0a977" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 13.834s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.508556] env[63531]: DEBUG nova.scheduler.client.report [None req-9b48ef40-e853-4488-bc73-2ca4984fb08c tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1000.560300] env[63531]: DEBUG nova.network.neutron [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1000.580105] env[63531]: DEBUG oslo_concurrency.lockutils [None req-dda659e3-a8a6-41cd-bdda-ec24d2209ba3 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "29dd6de3-2765-408c-acc0-da47e5e0a977" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.580305] env[63531]: DEBUG oslo_concurrency.lockutils [None req-dda659e3-a8a6-41cd-bdda-ec24d2209ba3 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "29dd6de3-2765-408c-acc0-da47e5e0a977" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.580484] env[63531]: DEBUG oslo_concurrency.lockutils [None req-dda659e3-a8a6-41cd-bdda-ec24d2209ba3 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "29dd6de3-2765-408c-acc0-da47e5e0a977-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.580674] env[63531]: DEBUG oslo_concurrency.lockutils [None req-dda659e3-a8a6-41cd-bdda-ec24d2209ba3 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "29dd6de3-2765-408c-acc0-da47e5e0a977-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.580848] env[63531]: DEBUG oslo_concurrency.lockutils [None req-dda659e3-a8a6-41cd-bdda-ec24d2209ba3 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "29dd6de3-2765-408c-acc0-da47e5e0a977-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.584954] env[63531]: INFO nova.compute.manager [None req-dda659e3-a8a6-41cd-bdda-ec24d2209ba3 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Terminating instance [ 1000.586860] env[63531]: DEBUG nova.compute.manager [None req-dda659e3-a8a6-41cd-bdda-ec24d2209ba3 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1000.587080] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-dda659e3-a8a6-41cd-bdda-ec24d2209ba3 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1000.587962] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bc4c64c-28fb-4ff5-bc36-48956ca7596a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.600487] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-dda659e3-a8a6-41cd-bdda-ec24d2209ba3 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1000.601319] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a24b691c-0486-4675-b318-8b05cbf11333 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.606638] env[63531]: DEBUG oslo_vmware.api [None req-0f17d244-380a-4408-853a-1a2f399662e9 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118385, 'name': ReconfigVM_Task, 'duration_secs': 0.649434} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.606916] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-0f17d244-380a-4408-853a-1a2f399662e9 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Reconfigured VM instance instance-00000055 to attach disk [datastore1] volume-3ac35683-e14e-4d02-8b0f-a5ea34354c7f/volume-3ac35683-e14e-4d02-8b0f-a5ea34354c7f.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1000.612942] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ab5568c6-6c96-4ad4-8e93-5c7e15998d04 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.625082] env[63531]: DEBUG oslo_vmware.api [None req-dda659e3-a8a6-41cd-bdda-ec24d2209ba3 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 1000.625082] env[63531]: value = "task-1118388" [ 1000.625082] env[63531]: _type = "Task" [ 1000.625082] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.632414] env[63531]: DEBUG oslo_vmware.api [None req-0f17d244-380a-4408-853a-1a2f399662e9 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 1000.632414] env[63531]: value = "task-1118389" [ 1000.632414] env[63531]: _type = "Task" [ 1000.632414] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.637917] env[63531]: DEBUG oslo_vmware.api [None req-dda659e3-a8a6-41cd-bdda-ec24d2209ba3 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118388, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.650846] env[63531]: DEBUG oslo_vmware.api [None req-0f17d244-380a-4408-853a-1a2f399662e9 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118389, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.690861] env[63531]: DEBUG nova.network.neutron [req-276d1c3b-47a9-44ce-964d-3f1040caea71 req-d296b6cb-0de3-4ba3-8e8e-c2782bba6176 service nova] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Updated VIF entry in instance network info cache for port 7433ce46-ff5a-4e78-bcc9-9dcf2264a01b. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1000.691275] env[63531]: DEBUG nova.network.neutron [req-276d1c3b-47a9-44ce-964d-3f1040caea71 req-d296b6cb-0de3-4ba3-8e8e-c2782bba6176 service nova] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Updating instance_info_cache with network_info: [{"id": "7433ce46-ff5a-4e78-bcc9-9dcf2264a01b", "address": "fa:16:3e:ad:77:b5", "network": {"id": "913045d9-e90b-4366-987c-7706b73a6147", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1696510389-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f71c7db0a3614995becc25d9a8dce787", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c297fe21-cd0b-4226-813b-a65d2358d034", "external-id": "nsx-vlan-transportzone-98", "segmentation_id": 98, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7433ce46-ff", "ovs_interfaceid": "7433ce46-ff5a-4e78-bcc9-9dcf2264a01b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1000.698548] env[63531]: DEBUG oslo_vmware.api [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118387, 'name': PowerOnVM_Task, 'duration_secs': 0.477133} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.698802] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1000.699038] env[63531]: INFO nova.compute.manager [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Took 8.03 seconds to spawn the instance on the hypervisor. [ 1000.699304] env[63531]: DEBUG nova.compute.manager [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1000.700060] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15bbacbf-266b-4d67-a01a-72d99fbfafbe {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.766576] env[63531]: DEBUG nova.network.neutron [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Updating instance_info_cache with network_info: [{"id": "40adf4b3-527a-4414-b2dc-a24a39b897c6", "address": "fa:16:3e:0d:67:83", "network": {"id": "4dd26082-3411-4d99-a78b-c76c15703506", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1190890625-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ee5fd84e1da04fd984f81578e4085850", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6046aec4-feda-4ef9-bf4a-800de1e0cd3b", "external-id": "nsx-vlan-transportzone-903", "segmentation_id": 903, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap40adf4b3-52", "ovs_interfaceid": "40adf4b3-527a-4414-b2dc-a24a39b897c6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.013388] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9b48ef40-e853-4488-bc73-2ca4984fb08c tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.830s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.016073] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0f3b984b-d279-4944-8a9e-cb9bdebbf475 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.212s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.016248] env[63531]: DEBUG nova.objects.instance [None req-0f3b984b-d279-4944-8a9e-cb9bdebbf475 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Lazy-loading 'resources' on Instance uuid baf2db4e-2cdb-47f5-9a5c-97a233a459c9 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1001.040353] env[63531]: INFO nova.scheduler.client.report [None req-9b48ef40-e853-4488-bc73-2ca4984fb08c tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Deleted allocations for instance 0fb3b687-62cc-4296-b4ed-537aba18c880 [ 1001.135149] env[63531]: DEBUG oslo_vmware.api [None req-dda659e3-a8a6-41cd-bdda-ec24d2209ba3 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118388, 'name': PowerOffVM_Task, 'duration_secs': 0.298212} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.135433] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-dda659e3-a8a6-41cd-bdda-ec24d2209ba3 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1001.135614] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-dda659e3-a8a6-41cd-bdda-ec24d2209ba3 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1001.135888] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5bb0f96b-2ff8-4d8f-b33a-91a89a56e53a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.146727] env[63531]: DEBUG oslo_vmware.api [None req-0f17d244-380a-4408-853a-1a2f399662e9 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118389, 'name': ReconfigVM_Task, 'duration_secs': 0.156477} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.147071] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-0f17d244-380a-4408-853a-1a2f399662e9 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244776', 'volume_id': '3ac35683-e14e-4d02-8b0f-a5ea34354c7f', 'name': 'volume-3ac35683-e14e-4d02-8b0f-a5ea34354c7f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '911f23f7-e320-48f3-87a9-0239b013feff', 'attached_at': '', 'detached_at': '', 'volume_id': '3ac35683-e14e-4d02-8b0f-a5ea34354c7f', 'serial': '3ac35683-e14e-4d02-8b0f-a5ea34354c7f'} {{(pid=63531) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1001.194239] env[63531]: DEBUG oslo_concurrency.lockutils [req-276d1c3b-47a9-44ce-964d-3f1040caea71 req-d296b6cb-0de3-4ba3-8e8e-c2782bba6176 service nova] Releasing lock "refresh_cache-ccf00b09-29fe-4a76-a8af-97774acd77de" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1001.194432] env[63531]: DEBUG nova.compute.manager [req-276d1c3b-47a9-44ce-964d-3f1040caea71 req-d296b6cb-0de3-4ba3-8e8e-c2782bba6176 service nova] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Received event network-changed-7433ce46-ff5a-4e78-bcc9-9dcf2264a01b {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1001.194686] env[63531]: DEBUG nova.compute.manager [req-276d1c3b-47a9-44ce-964d-3f1040caea71 req-d296b6cb-0de3-4ba3-8e8e-c2782bba6176 service nova] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Refreshing instance network info cache due to event network-changed-7433ce46-ff5a-4e78-bcc9-9dcf2264a01b. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1001.194949] env[63531]: DEBUG oslo_concurrency.lockutils [req-276d1c3b-47a9-44ce-964d-3f1040caea71 req-d296b6cb-0de3-4ba3-8e8e-c2782bba6176 service nova] Acquiring lock "refresh_cache-ccf00b09-29fe-4a76-a8af-97774acd77de" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1001.195142] env[63531]: DEBUG oslo_concurrency.lockutils [req-276d1c3b-47a9-44ce-964d-3f1040caea71 req-d296b6cb-0de3-4ba3-8e8e-c2782bba6176 service nova] Acquired lock "refresh_cache-ccf00b09-29fe-4a76-a8af-97774acd77de" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1001.195313] env[63531]: DEBUG nova.network.neutron [req-276d1c3b-47a9-44ce-964d-3f1040caea71 req-d296b6cb-0de3-4ba3-8e8e-c2782bba6176 service nova] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Refreshing network info cache for port 7433ce46-ff5a-4e78-bcc9-9dcf2264a01b {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1001.218594] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-dda659e3-a8a6-41cd-bdda-ec24d2209ba3 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1001.218594] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-dda659e3-a8a6-41cd-bdda-ec24d2209ba3 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Deleting contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1001.218594] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-dda659e3-a8a6-41cd-bdda-ec24d2209ba3 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Deleting the datastore file [datastore2] 29dd6de3-2765-408c-acc0-da47e5e0a977 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1001.220546] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4de4e387-68f9-4430-b1b8-4be7d24b9c9c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.222863] env[63531]: INFO nova.compute.manager [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Took 20.36 seconds to build instance. [ 1001.230279] env[63531]: DEBUG oslo_vmware.api [None req-dda659e3-a8a6-41cd-bdda-ec24d2209ba3 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 1001.230279] env[63531]: value = "task-1118391" [ 1001.230279] env[63531]: _type = "Task" [ 1001.230279] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.239898] env[63531]: DEBUG oslo_vmware.api [None req-dda659e3-a8a6-41cd-bdda-ec24d2209ba3 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118391, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.269599] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Releasing lock "refresh_cache-b7b2b7ac-2321-4cf3-b2ed-4877b893af63" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1001.269981] env[63531]: DEBUG nova.compute.manager [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Instance network_info: |[{"id": "40adf4b3-527a-4414-b2dc-a24a39b897c6", "address": "fa:16:3e:0d:67:83", "network": {"id": "4dd26082-3411-4d99-a78b-c76c15703506", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1190890625-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ee5fd84e1da04fd984f81578e4085850", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6046aec4-feda-4ef9-bf4a-800de1e0cd3b", "external-id": "nsx-vlan-transportzone-903", "segmentation_id": 903, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap40adf4b3-52", "ovs_interfaceid": "40adf4b3-527a-4414-b2dc-a24a39b897c6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1001.270416] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0d:67:83', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6046aec4-feda-4ef9-bf4a-800de1e0cd3b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '40adf4b3-527a-4414-b2dc-a24a39b897c6', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1001.278026] env[63531]: DEBUG oslo.service.loopingcall [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1001.278543] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1001.278782] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f4a76594-3419-4990-9796-e0e4e900adc1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.299635] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1001.299635] env[63531]: value = "task-1118392" [ 1001.299635] env[63531]: _type = "Task" [ 1001.299635] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.310217] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118392, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.547601] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9b48ef40-e853-4488-bc73-2ca4984fb08c tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Lock "0fb3b687-62cc-4296-b4ed-537aba18c880" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.385s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.725333] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9ec4ccee-868c-4a1c-8d60-a4437148f946 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "18057005-1ff1-4e3d-bccc-8c89755d6db7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.878s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.747576] env[63531]: DEBUG oslo_vmware.api [None req-dda659e3-a8a6-41cd-bdda-ec24d2209ba3 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118391, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.164706} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.747834] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-dda659e3-a8a6-41cd-bdda-ec24d2209ba3 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1001.748100] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-dda659e3-a8a6-41cd-bdda-ec24d2209ba3 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Deleted contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1001.748295] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-dda659e3-a8a6-41cd-bdda-ec24d2209ba3 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1001.748506] env[63531]: INFO nova.compute.manager [None req-dda659e3-a8a6-41cd-bdda-ec24d2209ba3 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1001.748760] env[63531]: DEBUG oslo.service.loopingcall [None req-dda659e3-a8a6-41cd-bdda-ec24d2209ba3 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1001.748996] env[63531]: DEBUG nova.compute.manager [-] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1001.749107] env[63531]: DEBUG nova.network.neutron [-] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1001.759593] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cd62d2d-a969-492d-acdf-f4191f461e77 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.774026] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4151d690-9eb0-463c-9ab4-b41c36140825 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.818083] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acd24933-c49b-4337-9011-316e4b6780fc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.826215] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118392, 'name': CreateVM_Task, 'duration_secs': 0.356706} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.828408] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1001.829338] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1001.829725] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1001.830175] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1001.831467] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a498048c-3cde-4fe9-9d3f-a58b28d048a1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.836328] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1ec8258f-2a95-4c39-980c-0ac49ca6d960 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.851482] env[63531]: DEBUG nova.compute.provider_tree [None req-0f3b984b-d279-4944-8a9e-cb9bdebbf475 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1001.852965] env[63531]: DEBUG oslo_vmware.api [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Waiting for the task: (returnval){ [ 1001.852965] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52f404a2-8e3a-b059-f499-57de82d617cb" [ 1001.852965] env[63531]: _type = "Task" [ 1001.852965] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.863113] env[63531]: DEBUG oslo_vmware.api [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52f404a2-8e3a-b059-f499-57de82d617cb, 'name': SearchDatastore_Task, 'duration_secs': 0.01357} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.864284] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1001.864658] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1001.865046] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1001.865313] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1001.865601] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1001.866206] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bc533afd-6fd6-441b-ad11-e2c00931b4d9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.878813] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1001.878813] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1001.879744] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2771cadc-927b-4819-ad89-6ce0de879d5b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.887583] env[63531]: DEBUG oslo_vmware.api [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Waiting for the task: (returnval){ [ 1001.887583] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]523c73c8-0f0d-c4f8-2a4f-d3e442436b25" [ 1001.887583] env[63531]: _type = "Task" [ 1001.887583] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.897125] env[63531]: DEBUG oslo_vmware.api [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]523c73c8-0f0d-c4f8-2a4f-d3e442436b25, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.902859] env[63531]: DEBUG nova.compute.manager [req-1ff3f745-c942-42d3-b0f6-07012f8fd31a req-097e6b99-3565-454c-8a10-d9c246398ea5 service nova] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Received event network-changed-af2485f6-5ce2-4a83-9d95-7bea85cebff2 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1001.903075] env[63531]: DEBUG nova.compute.manager [req-1ff3f745-c942-42d3-b0f6-07012f8fd31a req-097e6b99-3565-454c-8a10-d9c246398ea5 service nova] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Refreshing instance network info cache due to event network-changed-af2485f6-5ce2-4a83-9d95-7bea85cebff2. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1001.903301] env[63531]: DEBUG oslo_concurrency.lockutils [req-1ff3f745-c942-42d3-b0f6-07012f8fd31a req-097e6b99-3565-454c-8a10-d9c246398ea5 service nova] Acquiring lock "refresh_cache-18057005-1ff1-4e3d-bccc-8c89755d6db7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1001.903449] env[63531]: DEBUG oslo_concurrency.lockutils [req-1ff3f745-c942-42d3-b0f6-07012f8fd31a req-097e6b99-3565-454c-8a10-d9c246398ea5 service nova] Acquired lock "refresh_cache-18057005-1ff1-4e3d-bccc-8c89755d6db7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1001.903675] env[63531]: DEBUG nova.network.neutron [req-1ff3f745-c942-42d3-b0f6-07012f8fd31a req-097e6b99-3565-454c-8a10-d9c246398ea5 service nova] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Refreshing network info cache for port af2485f6-5ce2-4a83-9d95-7bea85cebff2 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1002.190500] env[63531]: DEBUG nova.network.neutron [req-276d1c3b-47a9-44ce-964d-3f1040caea71 req-d296b6cb-0de3-4ba3-8e8e-c2782bba6176 service nova] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Updated VIF entry in instance network info cache for port 7433ce46-ff5a-4e78-bcc9-9dcf2264a01b. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1002.190969] env[63531]: DEBUG nova.network.neutron [req-276d1c3b-47a9-44ce-964d-3f1040caea71 req-d296b6cb-0de3-4ba3-8e8e-c2782bba6176 service nova] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Updating instance_info_cache with network_info: [{"id": "7433ce46-ff5a-4e78-bcc9-9dcf2264a01b", "address": "fa:16:3e:ad:77:b5", "network": {"id": "913045d9-e90b-4366-987c-7706b73a6147", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1696510389-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f71c7db0a3614995becc25d9a8dce787", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c297fe21-cd0b-4226-813b-a65d2358d034", "external-id": "nsx-vlan-transportzone-98", "segmentation_id": 98, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7433ce46-ff", "ovs_interfaceid": "7433ce46-ff5a-4e78-bcc9-9dcf2264a01b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1002.192774] env[63531]: DEBUG nova.objects.instance [None req-0f17d244-380a-4408-853a-1a2f399662e9 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lazy-loading 'flavor' on Instance uuid 911f23f7-e320-48f3-87a9-0239b013feff {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1002.355493] env[63531]: DEBUG nova.scheduler.client.report [None req-0f3b984b-d279-4944-8a9e-cb9bdebbf475 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1002.399149] env[63531]: DEBUG oslo_vmware.api [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]523c73c8-0f0d-c4f8-2a4f-d3e442436b25, 'name': SearchDatastore_Task, 'duration_secs': 0.012291} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.399991] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0672ea61-bfa1-4bd5-bd08-cdc90a9a00e7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.408319] env[63531]: DEBUG oslo_vmware.api [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Waiting for the task: (returnval){ [ 1002.408319] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]526f49b7-2cfd-0914-b597-f31dfe1c5ef1" [ 1002.408319] env[63531]: _type = "Task" [ 1002.408319] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.417894] env[63531]: DEBUG oslo_vmware.api [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]526f49b7-2cfd-0914-b597-f31dfe1c5ef1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.695604] env[63531]: DEBUG oslo_concurrency.lockutils [req-276d1c3b-47a9-44ce-964d-3f1040caea71 req-d296b6cb-0de3-4ba3-8e8e-c2782bba6176 service nova] Releasing lock "refresh_cache-ccf00b09-29fe-4a76-a8af-97774acd77de" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.695941] env[63531]: DEBUG nova.compute.manager [req-276d1c3b-47a9-44ce-964d-3f1040caea71 req-d296b6cb-0de3-4ba3-8e8e-c2782bba6176 service nova] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Received event network-vif-plugged-40adf4b3-527a-4414-b2dc-a24a39b897c6 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1002.696221] env[63531]: DEBUG oslo_concurrency.lockutils [req-276d1c3b-47a9-44ce-964d-3f1040caea71 req-d296b6cb-0de3-4ba3-8e8e-c2782bba6176 service nova] Acquiring lock "b7b2b7ac-2321-4cf3-b2ed-4877b893af63-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.696465] env[63531]: DEBUG oslo_concurrency.lockutils [req-276d1c3b-47a9-44ce-964d-3f1040caea71 req-d296b6cb-0de3-4ba3-8e8e-c2782bba6176 service nova] Lock "b7b2b7ac-2321-4cf3-b2ed-4877b893af63-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.696642] env[63531]: DEBUG oslo_concurrency.lockutils [req-276d1c3b-47a9-44ce-964d-3f1040caea71 req-d296b6cb-0de3-4ba3-8e8e-c2782bba6176 service nova] Lock "b7b2b7ac-2321-4cf3-b2ed-4877b893af63-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.696839] env[63531]: DEBUG nova.compute.manager [req-276d1c3b-47a9-44ce-964d-3f1040caea71 req-d296b6cb-0de3-4ba3-8e8e-c2782bba6176 service nova] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] No waiting events found dispatching network-vif-plugged-40adf4b3-527a-4414-b2dc-a24a39b897c6 {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1002.697040] env[63531]: WARNING nova.compute.manager [req-276d1c3b-47a9-44ce-964d-3f1040caea71 req-d296b6cb-0de3-4ba3-8e8e-c2782bba6176 service nova] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Received unexpected event network-vif-plugged-40adf4b3-527a-4414-b2dc-a24a39b897c6 for instance with vm_state building and task_state spawning. [ 1002.697222] env[63531]: DEBUG nova.compute.manager [req-276d1c3b-47a9-44ce-964d-3f1040caea71 req-d296b6cb-0de3-4ba3-8e8e-c2782bba6176 service nova] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Received event network-changed-40adf4b3-527a-4414-b2dc-a24a39b897c6 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1002.697402] env[63531]: DEBUG nova.compute.manager [req-276d1c3b-47a9-44ce-964d-3f1040caea71 req-d296b6cb-0de3-4ba3-8e8e-c2782bba6176 service nova] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Refreshing instance network info cache due to event network-changed-40adf4b3-527a-4414-b2dc-a24a39b897c6. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1002.697596] env[63531]: DEBUG oslo_concurrency.lockutils [req-276d1c3b-47a9-44ce-964d-3f1040caea71 req-d296b6cb-0de3-4ba3-8e8e-c2782bba6176 service nova] Acquiring lock "refresh_cache-b7b2b7ac-2321-4cf3-b2ed-4877b893af63" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.697752] env[63531]: DEBUG oslo_concurrency.lockutils [req-276d1c3b-47a9-44ce-964d-3f1040caea71 req-d296b6cb-0de3-4ba3-8e8e-c2782bba6176 service nova] Acquired lock "refresh_cache-b7b2b7ac-2321-4cf3-b2ed-4877b893af63" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.697958] env[63531]: DEBUG nova.network.neutron [req-276d1c3b-47a9-44ce-964d-3f1040caea71 req-d296b6cb-0de3-4ba3-8e8e-c2782bba6176 service nova] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Refreshing network info cache for port 40adf4b3-527a-4414-b2dc-a24a39b897c6 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1002.704079] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0f17d244-380a-4408-853a-1a2f399662e9 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "911f23f7-e320-48f3-87a9-0239b013feff" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.299s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.860601] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0f3b984b-d279-4944-8a9e-cb9bdebbf475 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.845s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.862946] env[63531]: DEBUG oslo_concurrency.lockutils [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.302s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.864828] env[63531]: INFO nova.compute.claims [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1002.882121] env[63531]: INFO nova.scheduler.client.report [None req-0f3b984b-d279-4944-8a9e-cb9bdebbf475 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Deleted allocations for instance baf2db4e-2cdb-47f5-9a5c-97a233a459c9 [ 1002.920961] env[63531]: DEBUG oslo_vmware.api [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]526f49b7-2cfd-0914-b597-f31dfe1c5ef1, 'name': SearchDatastore_Task, 'duration_secs': 0.012143} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.923624] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.923906] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] b7b2b7ac-2321-4cf3-b2ed-4877b893af63/b7b2b7ac-2321-4cf3-b2ed-4877b893af63.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1002.924192] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-db48ea20-8901-408c-806a-7fddb813c570 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.933763] env[63531]: DEBUG oslo_vmware.api [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Waiting for the task: (returnval){ [ 1002.933763] env[63531]: value = "task-1118393" [ 1002.933763] env[63531]: _type = "Task" [ 1002.933763] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.943372] env[63531]: DEBUG oslo_vmware.api [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118393, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.997438] env[63531]: DEBUG nova.network.neutron [req-1ff3f745-c942-42d3-b0f6-07012f8fd31a req-097e6b99-3565-454c-8a10-d9c246398ea5 service nova] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Updated VIF entry in instance network info cache for port af2485f6-5ce2-4a83-9d95-7bea85cebff2. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1002.997930] env[63531]: DEBUG nova.network.neutron [req-1ff3f745-c942-42d3-b0f6-07012f8fd31a req-097e6b99-3565-454c-8a10-d9c246398ea5 service nova] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Updating instance_info_cache with network_info: [{"id": "af2485f6-5ce2-4a83-9d95-7bea85cebff2", "address": "fa:16:3e:f0:7d:92", "network": {"id": "c4f4c994-36ab-42ab-94f8-78f29cbafa50", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-994507698-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "57e9bf4ee0f1463fa67b0f8deff30fdc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "767a3a48-41d4-4a0c-961d-0024837f63bd", "external-id": "nsx-vlan-transportzone-141", "segmentation_id": 141, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaf2485f6-5c", "ovs_interfaceid": "af2485f6-5ce2-4a83-9d95-7bea85cebff2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1003.067133] env[63531]: DEBUG nova.network.neutron [-] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1003.261078] env[63531]: INFO nova.compute.manager [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Rebuilding instance [ 1003.314478] env[63531]: DEBUG nova.compute.manager [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1003.315418] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c66e9d3b-321d-4d88-a227-72d6929ade80 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.392739] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0f3b984b-d279-4944-8a9e-cb9bdebbf475 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Lock "baf2db4e-2cdb-47f5-9a5c-97a233a459c9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.162s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.444217] env[63531]: DEBUG oslo_vmware.api [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118393, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.467378} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.444570] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] b7b2b7ac-2321-4cf3-b2ed-4877b893af63/b7b2b7ac-2321-4cf3-b2ed-4877b893af63.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1003.444739] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1003.445352] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a0a19892-9c6e-4b6e-8b2e-700f5705a0fa {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.453430] env[63531]: DEBUG oslo_vmware.api [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Waiting for the task: (returnval){ [ 1003.453430] env[63531]: value = "task-1118394" [ 1003.453430] env[63531]: _type = "Task" [ 1003.453430] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.463606] env[63531]: DEBUG oslo_vmware.api [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118394, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.500812] env[63531]: DEBUG oslo_concurrency.lockutils [req-1ff3f745-c942-42d3-b0f6-07012f8fd31a req-097e6b99-3565-454c-8a10-d9c246398ea5 service nova] Releasing lock "refresh_cache-18057005-1ff1-4e3d-bccc-8c89755d6db7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1003.514744] env[63531]: DEBUG nova.network.neutron [req-276d1c3b-47a9-44ce-964d-3f1040caea71 req-d296b6cb-0de3-4ba3-8e8e-c2782bba6176 service nova] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Updated VIF entry in instance network info cache for port 40adf4b3-527a-4414-b2dc-a24a39b897c6. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1003.515163] env[63531]: DEBUG nova.network.neutron [req-276d1c3b-47a9-44ce-964d-3f1040caea71 req-d296b6cb-0de3-4ba3-8e8e-c2782bba6176 service nova] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Updating instance_info_cache with network_info: [{"id": "40adf4b3-527a-4414-b2dc-a24a39b897c6", "address": "fa:16:3e:0d:67:83", "network": {"id": "4dd26082-3411-4d99-a78b-c76c15703506", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1190890625-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ee5fd84e1da04fd984f81578e4085850", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6046aec4-feda-4ef9-bf4a-800de1e0cd3b", "external-id": "nsx-vlan-transportzone-903", "segmentation_id": 903, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap40adf4b3-52", "ovs_interfaceid": "40adf4b3-527a-4414-b2dc-a24a39b897c6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1003.569660] env[63531]: INFO nova.compute.manager [-] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Took 1.82 seconds to deallocate network for instance. [ 1003.832072] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1003.832362] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3cf21879-b79b-43a2-a8d9-975d24f7767e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.841034] env[63531]: DEBUG oslo_vmware.api [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 1003.841034] env[63531]: value = "task-1118395" [ 1003.841034] env[63531]: _type = "Task" [ 1003.841034] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.851035] env[63531]: DEBUG oslo_vmware.api [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118395, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.931894] env[63531]: DEBUG nova.compute.manager [req-ad357054-e4ea-404a-8d24-373d54d32ed1 req-ec7040d1-0db3-4d0d-9b7c-82e3eaa8b3d8 service nova] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Received event network-vif-deleted-262abfb7-7bcf-40e6-9ce3-b502de45b3b4 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1003.965656] env[63531]: DEBUG oslo_vmware.api [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118394, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072031} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.965900] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1003.966778] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbd81b01-adaf-4fbd-acc5-18e6caada5b4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.993240] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Reconfiguring VM instance instance-0000005d to attach disk [datastore2] b7b2b7ac-2321-4cf3-b2ed-4877b893af63/b7b2b7ac-2321-4cf3-b2ed-4877b893af63.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1003.996636] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e66efebf-bc01-47cc-bb83-698413feb01d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.017830] env[63531]: DEBUG oslo_concurrency.lockutils [req-276d1c3b-47a9-44ce-964d-3f1040caea71 req-d296b6cb-0de3-4ba3-8e8e-c2782bba6176 service nova] Releasing lock "refresh_cache-b7b2b7ac-2321-4cf3-b2ed-4877b893af63" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1004.021095] env[63531]: DEBUG oslo_vmware.api [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Waiting for the task: (returnval){ [ 1004.021095] env[63531]: value = "task-1118396" [ 1004.021095] env[63531]: _type = "Task" [ 1004.021095] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.034109] env[63531]: DEBUG oslo_vmware.api [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118396, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.076492] env[63531]: DEBUG oslo_concurrency.lockutils [None req-dda659e3-a8a6-41cd-bdda-ec24d2209ba3 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.114449] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73bf81dd-f8fb-463b-b657-6c513ff8aa77 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.123088] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41f5c313-0a65-40c0-bb08-4fbabbc53217 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.154898] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da215c34-99a9-4c4a-8dc4-7631a7189d98 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.163571] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5581e42-3dca-4f64-a5c3-50843628a954 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.177501] env[63531]: DEBUG nova.compute.provider_tree [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1004.352228] env[63531]: DEBUG oslo_vmware.api [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118395, 'name': PowerOffVM_Task, 'duration_secs': 0.241161} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.352525] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1004.408653] env[63531]: INFO nova.compute.manager [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Detaching volume 3ac35683-e14e-4d02-8b0f-a5ea34354c7f [ 1004.441986] env[63531]: INFO nova.virt.block_device [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Attempting to driver detach volume 3ac35683-e14e-4d02-8b0f-a5ea34354c7f from mountpoint /dev/sdb [ 1004.442239] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Volume detach. Driver type: vmdk {{(pid=63531) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1004.442446] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244776', 'volume_id': '3ac35683-e14e-4d02-8b0f-a5ea34354c7f', 'name': 'volume-3ac35683-e14e-4d02-8b0f-a5ea34354c7f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '911f23f7-e320-48f3-87a9-0239b013feff', 'attached_at': '', 'detached_at': '', 'volume_id': '3ac35683-e14e-4d02-8b0f-a5ea34354c7f', 'serial': '3ac35683-e14e-4d02-8b0f-a5ea34354c7f'} {{(pid=63531) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1004.443399] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-893545e3-e01a-41eb-926b-81f7a1ba6b32 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.466381] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f9a7079-52dc-4739-8c67-b6c21f6c292c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.473907] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bbc7bef-60a2-4f74-bbe5-a8d9465f18a3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.494749] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fd2a1df-b39b-419e-9669-a424c131ecb9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.511199] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] The volume has not been displaced from its original location: [datastore1] volume-3ac35683-e14e-4d02-8b0f-a5ea34354c7f/volume-3ac35683-e14e-4d02-8b0f-a5ea34354c7f.vmdk. No consolidation needed. {{(pid=63531) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1004.516454] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Reconfiguring VM instance instance-00000055 to detach disk 2001 {{(pid=63531) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1004.516736] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7c37bcb6-7be7-4221-8f27-96aeba158010 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.539411] env[63531]: DEBUG oslo_vmware.api [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118396, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.540697] env[63531]: DEBUG oslo_vmware.api [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 1004.540697] env[63531]: value = "task-1118397" [ 1004.540697] env[63531]: _type = "Task" [ 1004.540697] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.548494] env[63531]: DEBUG oslo_vmware.api [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118397, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.681065] env[63531]: DEBUG nova.scheduler.client.report [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1004.924071] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Acquiring lock "afa5f2c8-e519-4eb7-8e0e-ada9efb630e7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.924441] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Lock "afa5f2c8-e519-4eb7-8e0e-ada9efb630e7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.957909] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Acquiring lock "ba97ce22-ea4f-430b-a4c6-08204a568436" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.958239] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Lock "ba97ce22-ea4f-430b-a4c6-08204a568436" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.040866] env[63531]: DEBUG oslo_vmware.api [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118396, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.050022] env[63531]: DEBUG oslo_vmware.api [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118397, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.185781] env[63531]: DEBUG oslo_concurrency.lockutils [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.323s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.186741] env[63531]: DEBUG nova.compute.manager [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1005.190415] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4818e295-516e-436b-8abc-55b2e5232248 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.266s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.190677] env[63531]: DEBUG nova.objects.instance [None req-4818e295-516e-436b-8abc-55b2e5232248 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lazy-loading 'resources' on Instance uuid 3bf08a09-b6e6-4a58-8d35-7061a8e59afd {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1005.427254] env[63531]: DEBUG nova.compute.manager [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: afa5f2c8-e519-4eb7-8e0e-ada9efb630e7] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1005.460331] env[63531]: DEBUG nova.compute.manager [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: ba97ce22-ea4f-430b-a4c6-08204a568436] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1005.540937] env[63531]: DEBUG oslo_vmware.api [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118396, 'name': ReconfigVM_Task, 'duration_secs': 1.151506} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.541448] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Reconfigured VM instance instance-0000005d to attach disk [datastore2] b7b2b7ac-2321-4cf3-b2ed-4877b893af63/b7b2b7ac-2321-4cf3-b2ed-4877b893af63.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1005.544890] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b27e3f71-3e2d-4a32-8e54-c9542d423535 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.551968] env[63531]: DEBUG oslo_vmware.api [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118397, 'name': ReconfigVM_Task, 'duration_secs': 0.667003} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.553245] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Reconfigured VM instance instance-00000055 to detach disk 2001 {{(pid=63531) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1005.557854] env[63531]: DEBUG oslo_vmware.api [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Waiting for the task: (returnval){ [ 1005.557854] env[63531]: value = "task-1118398" [ 1005.557854] env[63531]: _type = "Task" [ 1005.557854] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.558397] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5ff70994-c25e-45ea-b37e-b3d8ecd77aab {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.579569] env[63531]: DEBUG oslo_vmware.api [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118398, 'name': Rename_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.581401] env[63531]: DEBUG oslo_vmware.api [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 1005.581401] env[63531]: value = "task-1118399" [ 1005.581401] env[63531]: _type = "Task" [ 1005.581401] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.593039] env[63531]: DEBUG oslo_vmware.api [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118399, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.693392] env[63531]: DEBUG nova.compute.utils [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1005.694953] env[63531]: DEBUG nova.compute.manager [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1005.695689] env[63531]: DEBUG nova.network.neutron [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1005.745908] env[63531]: DEBUG nova.policy [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '049cb89a769f41bd91d0e62784635cbd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c2e99349303e413b9f7896d449cb7dff', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 1005.888537] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e0cd544-4a00-4485-aa51-0ba60e4414f4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.897136] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1eb06b81-8205-4942-b2e1-fdf24a5f863f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.928024] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d706e8a-b737-4ba7-af00-095ccfa5cf84 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.940087] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3519fbf4-11ae-4a28-a395-2d57d4652098 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.946070] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.954218] env[63531]: DEBUG nova.compute.provider_tree [None req-4818e295-516e-436b-8abc-55b2e5232248 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1005.985148] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.007083] env[63531]: DEBUG nova.network.neutron [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Successfully created port: 91a7c60f-cbcb-4fe3-9bd0-69337882b7ce {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1006.078509] env[63531]: DEBUG oslo_vmware.api [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118398, 'name': Rename_Task, 'duration_secs': 0.141163} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.078798] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1006.079082] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-46f5f256-323d-44be-8722-be8b5660bed5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.086570] env[63531]: DEBUG oslo_vmware.api [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Waiting for the task: (returnval){ [ 1006.086570] env[63531]: value = "task-1118400" [ 1006.086570] env[63531]: _type = "Task" [ 1006.086570] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.091990] env[63531]: DEBUG oslo_vmware.api [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118399, 'name': ReconfigVM_Task, 'duration_secs': 0.139229} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.092574] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244776', 'volume_id': '3ac35683-e14e-4d02-8b0f-a5ea34354c7f', 'name': 'volume-3ac35683-e14e-4d02-8b0f-a5ea34354c7f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '911f23f7-e320-48f3-87a9-0239b013feff', 'attached_at': '', 'detached_at': '', 'volume_id': '3ac35683-e14e-4d02-8b0f-a5ea34354c7f', 'serial': '3ac35683-e14e-4d02-8b0f-a5ea34354c7f'} {{(pid=63531) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1006.097737] env[63531]: DEBUG oslo_vmware.api [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118400, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.201692] env[63531]: DEBUG nova.compute.manager [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1006.457077] env[63531]: DEBUG nova.scheduler.client.report [None req-4818e295-516e-436b-8abc-55b2e5232248 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1006.598914] env[63531]: DEBUG oslo_vmware.api [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118400, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.866794] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1006.867150] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1006.962737] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4818e295-516e-436b-8abc-55b2e5232248 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.772s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.965426] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6b6dc263-18a2-4a55-883c-76c6818d2fb1 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.936s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.965540] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6b6dc263-18a2-4a55-883c-76c6818d2fb1 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.967347] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.898s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.969527] env[63531]: INFO nova.compute.claims [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1006.985461] env[63531]: INFO nova.scheduler.client.report [None req-4818e295-516e-436b-8abc-55b2e5232248 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Deleted allocations for instance 3bf08a09-b6e6-4a58-8d35-7061a8e59afd [ 1006.987227] env[63531]: INFO nova.scheduler.client.report [None req-6b6dc263-18a2-4a55-883c-76c6818d2fb1 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Deleted allocations for instance 27c23b7c-a68f-43c9-a125-bd657feb3c5b [ 1007.097230] env[63531]: DEBUG oslo_vmware.api [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118400, 'name': PowerOnVM_Task, 'duration_secs': 0.52214} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.097455] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1007.097661] env[63531]: INFO nova.compute.manager [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Took 8.24 seconds to spawn the instance on the hypervisor. [ 1007.097847] env[63531]: DEBUG nova.compute.manager [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1007.098643] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dfaf6ed-020c-4ac7-b1cb-cadd075f9c3d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.145112] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1007.145589] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4c77455e-4ecb-493e-98bc-b1ca1778aac3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.154525] env[63531]: DEBUG oslo_vmware.api [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 1007.154525] env[63531]: value = "task-1118401" [ 1007.154525] env[63531]: _type = "Task" [ 1007.154525] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.162304] env[63531]: DEBUG oslo_vmware.api [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118401, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.211437] env[63531]: DEBUG nova.compute.manager [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1007.235674] env[63531]: DEBUG nova.virt.hardware [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1007.235956] env[63531]: DEBUG nova.virt.hardware [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1007.236140] env[63531]: DEBUG nova.virt.hardware [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1007.236355] env[63531]: DEBUG nova.virt.hardware [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1007.236524] env[63531]: DEBUG nova.virt.hardware [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1007.236685] env[63531]: DEBUG nova.virt.hardware [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1007.236900] env[63531]: DEBUG nova.virt.hardware [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1007.237087] env[63531]: DEBUG nova.virt.hardware [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1007.237269] env[63531]: DEBUG nova.virt.hardware [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1007.237441] env[63531]: DEBUG nova.virt.hardware [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1007.237620] env[63531]: DEBUG nova.virt.hardware [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1007.238525] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f04d5590-140b-47c8-bd82-4fa50338ba40 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.247603] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cb05426-55fb-4db2-9d18-130fce6cfc3a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.374138] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1007.374499] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Starting heal instance info cache {{(pid=63531) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1007.496498] env[63531]: DEBUG oslo_concurrency.lockutils [None req-4818e295-516e-436b-8abc-55b2e5232248 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "3bf08a09-b6e6-4a58-8d35-7061a8e59afd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.661s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.498012] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6b6dc263-18a2-4a55-883c-76c6818d2fb1 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "27c23b7c-a68f-43c9-a125-bd657feb3c5b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.532s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.515550] env[63531]: DEBUG nova.compute.manager [req-5e5ce8d1-24b1-4be6-b4ed-7b9741988ef8 req-eb68bee0-88cc-44cc-957f-c7756956333b service nova] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Received event network-vif-plugged-91a7c60f-cbcb-4fe3-9bd0-69337882b7ce {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1007.515767] env[63531]: DEBUG oslo_concurrency.lockutils [req-5e5ce8d1-24b1-4be6-b4ed-7b9741988ef8 req-eb68bee0-88cc-44cc-957f-c7756956333b service nova] Acquiring lock "5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.515982] env[63531]: DEBUG oslo_concurrency.lockutils [req-5e5ce8d1-24b1-4be6-b4ed-7b9741988ef8 req-eb68bee0-88cc-44cc-957f-c7756956333b service nova] Lock "5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.516172] env[63531]: DEBUG oslo_concurrency.lockutils [req-5e5ce8d1-24b1-4be6-b4ed-7b9741988ef8 req-eb68bee0-88cc-44cc-957f-c7756956333b service nova] Lock "5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.516347] env[63531]: DEBUG nova.compute.manager [req-5e5ce8d1-24b1-4be6-b4ed-7b9741988ef8 req-eb68bee0-88cc-44cc-957f-c7756956333b service nova] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] No waiting events found dispatching network-vif-plugged-91a7c60f-cbcb-4fe3-9bd0-69337882b7ce {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1007.516515] env[63531]: WARNING nova.compute.manager [req-5e5ce8d1-24b1-4be6-b4ed-7b9741988ef8 req-eb68bee0-88cc-44cc-957f-c7756956333b service nova] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Received unexpected event network-vif-plugged-91a7c60f-cbcb-4fe3-9bd0-69337882b7ce for instance with vm_state building and task_state spawning. [ 1007.615861] env[63531]: INFO nova.compute.manager [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Took 20.13 seconds to build instance. [ 1007.666090] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] VM already powered off {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1007.666356] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Volume detach. Driver type: vmdk {{(pid=63531) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1007.666563] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244776', 'volume_id': '3ac35683-e14e-4d02-8b0f-a5ea34354c7f', 'name': 'volume-3ac35683-e14e-4d02-8b0f-a5ea34354c7f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '911f23f7-e320-48f3-87a9-0239b013feff', 'attached_at': '', 'detached_at': '', 'volume_id': '3ac35683-e14e-4d02-8b0f-a5ea34354c7f', 'serial': '3ac35683-e14e-4d02-8b0f-a5ea34354c7f'} {{(pid=63531) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1007.667656] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca8cbac1-2c87-42b3-a892-66afaa5d9f9e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.688255] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1016d86-7f52-4be4-8fd5-c803949203bf {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.692941] env[63531]: DEBUG nova.network.neutron [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Successfully updated port: 91a7c60f-cbcb-4fe3-9bd0-69337882b7ce {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1007.696437] env[63531]: WARNING nova.virt.vmwareapi.driver [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] The volume None does not exist!: nova.exception.DiskNotFound: Unable to find volume [ 1007.696717] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1007.697480] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f9f1e36-3075-42b9-8857-ffa22dfa7d12 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.705983] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1007.706325] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1966b454-05e8-4ba5-a095-d0a7122cc2bc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.785038] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1007.786372] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1007.786372] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Deleting the datastore file [datastore1] 911f23f7-e320-48f3-87a9-0239b013feff {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1007.786372] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-da21f6f8-a4d7-40e6-b2c8-8ae807d9c83f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.792780] env[63531]: DEBUG oslo_vmware.api [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 1007.792780] env[63531]: value = "task-1118403" [ 1007.792780] env[63531]: _type = "Task" [ 1007.792780] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.800877] env[63531]: DEBUG oslo_vmware.api [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118403, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.117941] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f5c98d1e-23c3-4c12-b970-76a2a1a60b9c tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lock "b7b2b7ac-2321-4cf3-b2ed-4877b893af63" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.646s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.158409] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c92ab035-c0cf-4919-a91d-8030aa96c400 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.166626] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07319d55-9f15-494e-9faf-0ec834c65f10 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.196847] env[63531]: DEBUG oslo_concurrency.lockutils [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "refresh_cache-5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.197021] env[63531]: DEBUG oslo_concurrency.lockutils [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquired lock "refresh_cache-5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.197161] env[63531]: DEBUG nova.network.neutron [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1008.198950] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-758c14c9-5ab6-4243-aa7b-1b857ac628f9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.207718] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4e8edb7-6cd4-40bc-9ecb-a4500a3e640f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.222110] env[63531]: DEBUG nova.compute.provider_tree [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1008.303651] env[63531]: DEBUG oslo_vmware.api [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118403, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.1396} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.303924] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1008.304131] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1008.304322] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1008.359847] env[63531]: DEBUG nova.compute.manager [req-e730e796-35c3-4983-9458-7e2bde14138e req-b4ce11fc-0d38-4ca4-805f-49f5e716c895 service nova] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Received event network-changed-40adf4b3-527a-4414-b2dc-a24a39b897c6 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1008.359962] env[63531]: DEBUG nova.compute.manager [req-e730e796-35c3-4983-9458-7e2bde14138e req-b4ce11fc-0d38-4ca4-805f-49f5e716c895 service nova] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Refreshing instance network info cache due to event network-changed-40adf4b3-527a-4414-b2dc-a24a39b897c6. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1008.360206] env[63531]: DEBUG oslo_concurrency.lockutils [req-e730e796-35c3-4983-9458-7e2bde14138e req-b4ce11fc-0d38-4ca4-805f-49f5e716c895 service nova] Acquiring lock "refresh_cache-b7b2b7ac-2321-4cf3-b2ed-4877b893af63" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.360356] env[63531]: DEBUG oslo_concurrency.lockutils [req-e730e796-35c3-4983-9458-7e2bde14138e req-b4ce11fc-0d38-4ca4-805f-49f5e716c895 service nova] Acquired lock "refresh_cache-b7b2b7ac-2321-4cf3-b2ed-4877b893af63" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.360523] env[63531]: DEBUG nova.network.neutron [req-e730e796-35c3-4983-9458-7e2bde14138e req-b4ce11fc-0d38-4ca4-805f-49f5e716c895 service nova] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Refreshing network info cache for port 40adf4b3-527a-4414-b2dc-a24a39b897c6 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1008.379029] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Didn't find any instances for network info cache update. {{(pid=63531) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1008.379206] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1008.379377] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1008.379531] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1008.379683] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1008.379827] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1008.379976] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1008.380123] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63531) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1008.380291] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager.update_available_resource {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1008.475330] env[63531]: DEBUG oslo_concurrency.lockutils [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "b6db77e6-4447-4b22-93d7-265ff0ada0bd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.475788] env[63531]: DEBUG oslo_concurrency.lockutils [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "b6db77e6-4447-4b22-93d7-265ff0ada0bd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.724982] env[63531]: DEBUG nova.scheduler.client.report [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1008.754520] env[63531]: DEBUG nova.network.neutron [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1008.809633] env[63531]: INFO nova.virt.block_device [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Booting with volume 3ac35683-e14e-4d02-8b0f-a5ea34354c7f at /dev/sdb [ 1008.845765] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bada68a4-84cf-4071-899b-e82c4b114ab3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.855574] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0149010b-0367-424e-b782-f115c4adf5e0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.891860] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.892425] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3e4822d4-1575-4aab-8333-a6afb9b3cc46 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.907026] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-890d6443-c6cc-409f-8806-8d3b6d601593 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.943329] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5f4b443-d21b-4db6-8ed5-0a0fc9e7fb0e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.953047] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e893015a-7467-456d-844d-10d179248859 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.972376] env[63531]: DEBUG nova.virt.block_device [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Updating existing volume attachment record: bfb4f39b-4e05-4389-9664-3531fd144dd7 {{(pid=63531) _volume_attach /opt/stack/nova/nova/virt/block_device.py:679}} [ 1008.975915] env[63531]: DEBUG nova.network.neutron [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Updating instance_info_cache with network_info: [{"id": "91a7c60f-cbcb-4fe3-9bd0-69337882b7ce", "address": "fa:16:3e:46:61:c6", "network": {"id": "bf99b8cb-9efb-4a74-9db1-0f5f0875f26a", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-998205595-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2e99349303e413b9f7896d449cb7dff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc6e6fe1-c4f5-4389-a49f-0978060eebb4", "external-id": "nsx-vlan-transportzone-829", "segmentation_id": 829, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91a7c60f-cb", "ovs_interfaceid": "91a7c60f-cbcb-4fe3-9bd0-69337882b7ce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1008.979526] env[63531]: DEBUG nova.compute.manager [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b6db77e6-4447-4b22-93d7-265ff0ada0bd] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1009.202641] env[63531]: DEBUG nova.network.neutron [req-e730e796-35c3-4983-9458-7e2bde14138e req-b4ce11fc-0d38-4ca4-805f-49f5e716c895 service nova] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Updated VIF entry in instance network info cache for port 40adf4b3-527a-4414-b2dc-a24a39b897c6. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1009.203053] env[63531]: DEBUG nova.network.neutron [req-e730e796-35c3-4983-9458-7e2bde14138e req-b4ce11fc-0d38-4ca4-805f-49f5e716c895 service nova] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Updating instance_info_cache with network_info: [{"id": "40adf4b3-527a-4414-b2dc-a24a39b897c6", "address": "fa:16:3e:0d:67:83", "network": {"id": "4dd26082-3411-4d99-a78b-c76c15703506", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1190890625-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.251", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ee5fd84e1da04fd984f81578e4085850", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6046aec4-feda-4ef9-bf4a-800de1e0cd3b", "external-id": "nsx-vlan-transportzone-903", "segmentation_id": 903, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap40adf4b3-52", "ovs_interfaceid": "40adf4b3-527a-4414-b2dc-a24a39b897c6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1009.230457] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.263s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.230977] env[63531]: DEBUG nova.compute.manager [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1009.233463] env[63531]: DEBUG oslo_concurrency.lockutils [None req-dda659e3-a8a6-41cd-bdda-ec24d2209ba3 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.157s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.233659] env[63531]: DEBUG oslo_concurrency.lockutils [None req-dda659e3-a8a6-41cd-bdda-ec24d2209ba3 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.235640] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.290s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.237454] env[63531]: INFO nova.compute.claims [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: afa5f2c8-e519-4eb7-8e0e-ada9efb630e7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1009.260216] env[63531]: INFO nova.scheduler.client.report [None req-dda659e3-a8a6-41cd-bdda-ec24d2209ba3 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Deleted allocations for instance 29dd6de3-2765-408c-acc0-da47e5e0a977 [ 1009.487389] env[63531]: DEBUG oslo_concurrency.lockutils [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Releasing lock "refresh_cache-5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.487789] env[63531]: DEBUG nova.compute.manager [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Instance network_info: |[{"id": "91a7c60f-cbcb-4fe3-9bd0-69337882b7ce", "address": "fa:16:3e:46:61:c6", "network": {"id": "bf99b8cb-9efb-4a74-9db1-0f5f0875f26a", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-998205595-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2e99349303e413b9f7896d449cb7dff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc6e6fe1-c4f5-4389-a49f-0978060eebb4", "external-id": "nsx-vlan-transportzone-829", "segmentation_id": 829, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91a7c60f-cb", "ovs_interfaceid": "91a7c60f-cbcb-4fe3-9bd0-69337882b7ce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1009.490749] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:46:61:c6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bc6e6fe1-c4f5-4389-a49f-0978060eebb4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '91a7c60f-cbcb-4fe3-9bd0-69337882b7ce', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1009.498405] env[63531]: DEBUG oslo.service.loopingcall [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1009.500842] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1009.501896] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2261bb58-dee1-4ca5-8544-98c9fbdcd4ff {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.515898] env[63531]: DEBUG oslo_concurrency.lockutils [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.523396] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1009.523396] env[63531]: value = "task-1118404" [ 1009.523396] env[63531]: _type = "Task" [ 1009.523396] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.534659] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118404, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.548608] env[63531]: DEBUG nova.compute.manager [req-4157b817-76cf-48e1-a2ea-0dca3db5ad22 req-62316e3c-ea51-4e9b-bf77-b71d571a7fc8 service nova] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Received event network-changed-91a7c60f-cbcb-4fe3-9bd0-69337882b7ce {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1009.548806] env[63531]: DEBUG nova.compute.manager [req-4157b817-76cf-48e1-a2ea-0dca3db5ad22 req-62316e3c-ea51-4e9b-bf77-b71d571a7fc8 service nova] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Refreshing instance network info cache due to event network-changed-91a7c60f-cbcb-4fe3-9bd0-69337882b7ce. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1009.549041] env[63531]: DEBUG oslo_concurrency.lockutils [req-4157b817-76cf-48e1-a2ea-0dca3db5ad22 req-62316e3c-ea51-4e9b-bf77-b71d571a7fc8 service nova] Acquiring lock "refresh_cache-5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1009.549193] env[63531]: DEBUG oslo_concurrency.lockutils [req-4157b817-76cf-48e1-a2ea-0dca3db5ad22 req-62316e3c-ea51-4e9b-bf77-b71d571a7fc8 service nova] Acquired lock "refresh_cache-5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1009.549363] env[63531]: DEBUG nova.network.neutron [req-4157b817-76cf-48e1-a2ea-0dca3db5ad22 req-62316e3c-ea51-4e9b-bf77-b71d571a7fc8 service nova] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Refreshing network info cache for port 91a7c60f-cbcb-4fe3-9bd0-69337882b7ce {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1009.705679] env[63531]: DEBUG oslo_concurrency.lockutils [req-e730e796-35c3-4983-9458-7e2bde14138e req-b4ce11fc-0d38-4ca4-805f-49f5e716c895 service nova] Releasing lock "refresh_cache-b7b2b7ac-2321-4cf3-b2ed-4877b893af63" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.741695] env[63531]: DEBUG nova.compute.utils [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1009.745653] env[63531]: DEBUG nova.compute.manager [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1009.745653] env[63531]: DEBUG nova.network.neutron [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1009.766700] env[63531]: DEBUG oslo_concurrency.lockutils [None req-dda659e3-a8a6-41cd-bdda-ec24d2209ba3 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "29dd6de3-2765-408c-acc0-da47e5e0a977" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.186s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.791755] env[63531]: DEBUG nova.policy [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0372c22d1e9e4f66a0417643dab4a2ec', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '94a95667c00947ea9183a6307c569c90', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 1010.037552] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118404, 'name': CreateVM_Task, 'duration_secs': 0.391646} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.037803] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1010.038506] env[63531]: DEBUG oslo_concurrency.lockutils [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1010.038747] env[63531]: DEBUG oslo_concurrency.lockutils [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.039182] env[63531]: DEBUG oslo_concurrency.lockutils [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1010.039413] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-57ef72c1-778e-4288-8649-327d07e94ae6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.047319] env[63531]: DEBUG oslo_vmware.api [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 1010.047319] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5295dd02-90a8-4331-cd08-07785a3c2ab1" [ 1010.047319] env[63531]: _type = "Task" [ 1010.047319] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.067695] env[63531]: DEBUG oslo_vmware.api [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5295dd02-90a8-4331-cd08-07785a3c2ab1, 'name': SearchDatastore_Task, 'duration_secs': 0.010812} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.068026] env[63531]: DEBUG oslo_concurrency.lockutils [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1010.068346] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1010.068674] env[63531]: DEBUG oslo_concurrency.lockutils [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1010.069173] env[63531]: DEBUG oslo_concurrency.lockutils [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.069402] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1010.070051] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bcfd7423-f945-48b9-9ec2-4f0f581334ee {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.081076] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1010.081344] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1010.082506] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-381b5b97-c7e6-48f7-b109-4fe358079adf {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.089586] env[63531]: DEBUG oslo_vmware.api [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 1010.089586] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52719b6b-440d-7940-eead-bf0bf95f1191" [ 1010.089586] env[63531]: _type = "Task" [ 1010.089586] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.100524] env[63531]: DEBUG oslo_vmware.api [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52719b6b-440d-7940-eead-bf0bf95f1191, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.165425] env[63531]: DEBUG nova.network.neutron [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Successfully created port: 03e99378-6a13-403e-9cb4-856abf141914 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1010.246101] env[63531]: DEBUG nova.compute.manager [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1010.458738] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f58748b-0819-4c94-a5df-6fa91e9bc1da {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.467121] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8ed1bad-dfdf-4d19-b702-782deb93db42 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.498316] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-570dca8a-20f9-4909-af7a-058ff5d424fa {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.506345] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04a04b82-756b-4b52-b63b-6b766bbd9544 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.512136] env[63531]: DEBUG nova.network.neutron [req-4157b817-76cf-48e1-a2ea-0dca3db5ad22 req-62316e3c-ea51-4e9b-bf77-b71d571a7fc8 service nova] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Updated VIF entry in instance network info cache for port 91a7c60f-cbcb-4fe3-9bd0-69337882b7ce. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1010.512479] env[63531]: DEBUG nova.network.neutron [req-4157b817-76cf-48e1-a2ea-0dca3db5ad22 req-62316e3c-ea51-4e9b-bf77-b71d571a7fc8 service nova] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Updating instance_info_cache with network_info: [{"id": "91a7c60f-cbcb-4fe3-9bd0-69337882b7ce", "address": "fa:16:3e:46:61:c6", "network": {"id": "bf99b8cb-9efb-4a74-9db1-0f5f0875f26a", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-998205595-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2e99349303e413b9f7896d449cb7dff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc6e6fe1-c4f5-4389-a49f-0978060eebb4", "external-id": "nsx-vlan-transportzone-829", "segmentation_id": 829, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91a7c60f-cb", "ovs_interfaceid": "91a7c60f-cbcb-4fe3-9bd0-69337882b7ce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1010.524631] env[63531]: DEBUG nova.compute.provider_tree [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1010.603715] env[63531]: DEBUG oslo_vmware.api [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52719b6b-440d-7940-eead-bf0bf95f1191, 'name': SearchDatastore_Task, 'duration_secs': 0.009922} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.606020] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-757e9103-24c6-4b7f-b3c0-3423659f6d56 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.612125] env[63531]: DEBUG oslo_vmware.api [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 1010.612125] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52fc1793-f79c-089c-53e1-0eb95ab0a538" [ 1010.612125] env[63531]: _type = "Task" [ 1010.612125] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.621460] env[63531]: DEBUG oslo_vmware.api [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52fc1793-f79c-089c-53e1-0eb95ab0a538, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.015656] env[63531]: DEBUG oslo_concurrency.lockutils [req-4157b817-76cf-48e1-a2ea-0dca3db5ad22 req-62316e3c-ea51-4e9b-bf77-b71d571a7fc8 service nova] Releasing lock "refresh_cache-5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1011.027977] env[63531]: DEBUG nova.scheduler.client.report [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1011.113533] env[63531]: DEBUG nova.virt.hardware [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1011.113790] env[63531]: DEBUG nova.virt.hardware [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1011.113952] env[63531]: DEBUG nova.virt.hardware [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1011.114158] env[63531]: DEBUG nova.virt.hardware [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1011.114380] env[63531]: DEBUG nova.virt.hardware [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1011.114691] env[63531]: DEBUG nova.virt.hardware [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1011.114939] env[63531]: DEBUG nova.virt.hardware [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1011.115137] env[63531]: DEBUG nova.virt.hardware [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1011.115318] env[63531]: DEBUG nova.virt.hardware [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1011.115495] env[63531]: DEBUG nova.virt.hardware [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1011.115676] env[63531]: DEBUG nova.virt.hardware [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1011.116701] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdb0673c-b940-4dd1-b82c-6747861d0c45 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.131292] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0bdfda7-dd99-4826-aacf-4e5c54fd8f7c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.135408] env[63531]: DEBUG oslo_vmware.api [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52fc1793-f79c-089c-53e1-0eb95ab0a538, 'name': SearchDatastore_Task, 'duration_secs': 0.010566} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.135675] env[63531]: DEBUG oslo_concurrency.lockutils [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1011.135944] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7/5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1011.136651] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f4ce9921-7da5-4ebe-8650-ebd93c892125 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.146505] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:71:30:b9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'db068f71-08cc-42d4-8ab6-17134c1585e5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd3f47a79-f699-4d63-9412-724fe1f15913', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1011.154093] env[63531]: DEBUG oslo.service.loopingcall [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1011.155688] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1011.156057] env[63531]: DEBUG oslo_vmware.api [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 1011.156057] env[63531]: value = "task-1118405" [ 1011.156057] env[63531]: _type = "Task" [ 1011.156057] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.156301] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1c68e296-7589-420e-b79a-441f885b7e8d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.180818] env[63531]: DEBUG oslo_vmware.api [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118405, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.182247] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1011.182247] env[63531]: value = "task-1118406" [ 1011.182247] env[63531]: _type = "Task" [ 1011.182247] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.190040] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118406, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.266905] env[63531]: DEBUG nova.compute.manager [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1011.288299] env[63531]: DEBUG nova.virt.hardware [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1011.288605] env[63531]: DEBUG nova.virt.hardware [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1011.288801] env[63531]: DEBUG nova.virt.hardware [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1011.289166] env[63531]: DEBUG nova.virt.hardware [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1011.289399] env[63531]: DEBUG nova.virt.hardware [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1011.289599] env[63531]: DEBUG nova.virt.hardware [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1011.289877] env[63531]: DEBUG nova.virt.hardware [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1011.290110] env[63531]: DEBUG nova.virt.hardware [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1011.290341] env[63531]: DEBUG nova.virt.hardware [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1011.290527] env[63531]: DEBUG nova.virt.hardware [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1011.290714] env[63531]: DEBUG nova.virt.hardware [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1011.291699] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c24b6e99-7082-40e0-a0fd-7ed35f1d737d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.300400] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db180f01-5f73-4433-a8b8-463fb59f1a4d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.362602] env[63531]: DEBUG oslo_concurrency.lockutils [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "96c2910b-1e25-4053-9ad9-c155402ef0ac" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.362855] env[63531]: DEBUG oslo_concurrency.lockutils [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "96c2910b-1e25-4053-9ad9-c155402ef0ac" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.532701] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.297s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.533461] env[63531]: DEBUG nova.compute.manager [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: afa5f2c8-e519-4eb7-8e0e-ada9efb630e7] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1011.536338] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.551s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.537799] env[63531]: INFO nova.compute.claims [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: ba97ce22-ea4f-430b-a4c6-08204a568436] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1011.682081] env[63531]: DEBUG oslo_vmware.api [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118405, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.691426] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118406, 'name': CreateVM_Task, 'duration_secs': 0.379841} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.691600] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1011.692273] env[63531]: DEBUG oslo_concurrency.lockutils [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1011.692451] env[63531]: DEBUG oslo_concurrency.lockutils [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.692792] env[63531]: DEBUG oslo_concurrency.lockutils [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1011.693078] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d5332360-b1a8-4a37-ac97-e8beb62e6e6a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.698241] env[63531]: DEBUG oslo_vmware.api [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 1011.698241] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52571904-5f12-d343-68af-609dae311ee2" [ 1011.698241] env[63531]: _type = "Task" [ 1011.698241] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.706916] env[63531]: DEBUG oslo_vmware.api [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52571904-5f12-d343-68af-609dae311ee2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.865657] env[63531]: DEBUG nova.compute.manager [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1011.912109] env[63531]: DEBUG nova.compute.manager [req-b6b64f4a-c5dc-418b-8f02-86ea2e477766 req-46e00ff5-e257-4849-8a9a-31836640455a service nova] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Received event network-vif-plugged-03e99378-6a13-403e-9cb4-856abf141914 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1011.912351] env[63531]: DEBUG oslo_concurrency.lockutils [req-b6b64f4a-c5dc-418b-8f02-86ea2e477766 req-46e00ff5-e257-4849-8a9a-31836640455a service nova] Acquiring lock "3fe7a7a8-fc43-4560-b101-227907c91fd0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.912569] env[63531]: DEBUG oslo_concurrency.lockutils [req-b6b64f4a-c5dc-418b-8f02-86ea2e477766 req-46e00ff5-e257-4849-8a9a-31836640455a service nova] Lock "3fe7a7a8-fc43-4560-b101-227907c91fd0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.912784] env[63531]: DEBUG oslo_concurrency.lockutils [req-b6b64f4a-c5dc-418b-8f02-86ea2e477766 req-46e00ff5-e257-4849-8a9a-31836640455a service nova] Lock "3fe7a7a8-fc43-4560-b101-227907c91fd0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.912919] env[63531]: DEBUG nova.compute.manager [req-b6b64f4a-c5dc-418b-8f02-86ea2e477766 req-46e00ff5-e257-4849-8a9a-31836640455a service nova] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] No waiting events found dispatching network-vif-plugged-03e99378-6a13-403e-9cb4-856abf141914 {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1011.913181] env[63531]: WARNING nova.compute.manager [req-b6b64f4a-c5dc-418b-8f02-86ea2e477766 req-46e00ff5-e257-4849-8a9a-31836640455a service nova] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Received unexpected event network-vif-plugged-03e99378-6a13-403e-9cb4-856abf141914 for instance with vm_state building and task_state spawning. [ 1012.042595] env[63531]: DEBUG nova.compute.utils [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1012.045915] env[63531]: DEBUG nova.compute.manager [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: afa5f2c8-e519-4eb7-8e0e-ada9efb630e7] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1012.046035] env[63531]: DEBUG nova.network.neutron [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: afa5f2c8-e519-4eb7-8e0e-ada9efb630e7] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1012.083114] env[63531]: DEBUG nova.network.neutron [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Successfully updated port: 03e99378-6a13-403e-9cb4-856abf141914 {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1012.100903] env[63531]: DEBUG nova.policy [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4186052665ce457b99e1fee971be05b4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4ab5ee56531c4e638f8b695b9e5d082f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 1012.184012] env[63531]: DEBUG oslo_vmware.api [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118405, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.551125} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.184455] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7/5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1012.184842] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1012.184946] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-135b2c3a-f639-4e15-894e-50c08a4d16b7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.192942] env[63531]: DEBUG oslo_vmware.api [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 1012.192942] env[63531]: value = "task-1118407" [ 1012.192942] env[63531]: _type = "Task" [ 1012.192942] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.208188] env[63531]: DEBUG oslo_vmware.api [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118407, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.215288] env[63531]: DEBUG oslo_vmware.api [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52571904-5f12-d343-68af-609dae311ee2, 'name': SearchDatastore_Task, 'duration_secs': 0.008546} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.215608] env[63531]: DEBUG oslo_concurrency.lockutils [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1012.215843] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1012.216409] env[63531]: DEBUG oslo_concurrency.lockutils [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1012.216409] env[63531]: DEBUG oslo_concurrency.lockutils [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.216582] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1012.216878] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3d584880-d9b3-4073-a66b-52f607f5e5c7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.226231] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1012.226490] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1012.227265] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-87054512-733e-4c63-bede-8c30ff2ed13c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.233533] env[63531]: DEBUG oslo_vmware.api [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 1012.233533] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52f25d07-025b-b069-9154-081b21bec683" [ 1012.233533] env[63531]: _type = "Task" [ 1012.233533] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.242488] env[63531]: DEBUG oslo_vmware.api [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52f25d07-025b-b069-9154-081b21bec683, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.388989] env[63531]: DEBUG oslo_concurrency.lockutils [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1012.429479] env[63531]: DEBUG nova.network.neutron [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: afa5f2c8-e519-4eb7-8e0e-ada9efb630e7] Successfully created port: b320eed5-1bc8-4774-9abd-e8cfcd640e56 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1012.546762] env[63531]: DEBUG nova.compute.manager [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: afa5f2c8-e519-4eb7-8e0e-ada9efb630e7] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1012.586296] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "refresh_cache-3fe7a7a8-fc43-4560-b101-227907c91fd0" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1012.586449] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquired lock "refresh_cache-3fe7a7a8-fc43-4560-b101-227907c91fd0" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.586600] env[63531]: DEBUG nova.network.neutron [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1012.705138] env[63531]: DEBUG oslo_vmware.api [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118407, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072397} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.705435] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1012.706262] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b10fae2-0220-4a32-8364-4fd6a069b82f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.731877] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Reconfiguring VM instance instance-0000005e to attach disk [datastore2] 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7/5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1012.734938] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c23553e5-06ec-4ac1-b1ab-b46c525cb243 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.758479] env[63531]: DEBUG oslo_vmware.api [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52f25d07-025b-b069-9154-081b21bec683, 'name': SearchDatastore_Task, 'duration_secs': 0.01007} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.762616] env[63531]: DEBUG oslo_vmware.api [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 1012.762616] env[63531]: value = "task-1118408" [ 1012.762616] env[63531]: _type = "Task" [ 1012.762616] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.763026] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7530f61b-48f7-449f-bb3e-f493a5bb6f94 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.771522] env[63531]: DEBUG oslo_vmware.api [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 1012.771522] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52aab405-0232-fcae-4219-50a5bd72cd64" [ 1012.771522] env[63531]: _type = "Task" [ 1012.771522] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.774781] env[63531]: DEBUG oslo_vmware.api [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118408, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.778869] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adcc6ede-5e56-4db6-8bba-f7215af333b3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.786887] env[63531]: DEBUG oslo_vmware.api [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52aab405-0232-fcae-4219-50a5bd72cd64, 'name': SearchDatastore_Task, 'duration_secs': 0.01211} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.788802] env[63531]: DEBUG oslo_concurrency.lockutils [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1012.789119] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 911f23f7-e320-48f3-87a9-0239b013feff/911f23f7-e320-48f3-87a9-0239b013feff.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1012.789422] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6deeb96c-2537-4ba1-a319-90331ec6956c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.791864] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-130cf66b-2f86-4992-9655-8f46f7fa5d1f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.801974] env[63531]: DEBUG oslo_vmware.api [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 1012.801974] env[63531]: value = "task-1118409" [ 1012.801974] env[63531]: _type = "Task" [ 1012.801974] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.831696] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73845ddc-f7f1-4faa-865c-181b919ffb62 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.845433] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a094306a-07e8-4f7c-95ad-df273b9d0bb3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.848968] env[63531]: DEBUG oslo_vmware.api [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118409, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.859503] env[63531]: DEBUG nova.compute.provider_tree [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1013.133701] env[63531]: DEBUG nova.network.neutron [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1013.277586] env[63531]: DEBUG oslo_vmware.api [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118408, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.316509] env[63531]: DEBUG nova.network.neutron [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Updating instance_info_cache with network_info: [{"id": "03e99378-6a13-403e-9cb4-856abf141914", "address": "fa:16:3e:3d:57:da", "network": {"id": "ea2d47d3-a502-4afd-9bc5-ee4a37017901", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1361282648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94a95667c00947ea9183a6307c569c90", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6c45fd-e930-495a-9cb7-df84eda443b1", "external-id": "nsx-vlan-transportzone-407", "segmentation_id": 407, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03e99378-6a", "ovs_interfaceid": "03e99378-6a13-403e-9cb4-856abf141914", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1013.340106] env[63531]: DEBUG oslo_vmware.api [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118409, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.462977} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.340296] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 911f23f7-e320-48f3-87a9-0239b013feff/911f23f7-e320-48f3-87a9-0239b013feff.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1013.340550] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1013.340938] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f53a4188-6793-4ed9-8d74-fb97f4c281c9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.348655] env[63531]: DEBUG oslo_vmware.api [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 1013.348655] env[63531]: value = "task-1118410" [ 1013.348655] env[63531]: _type = "Task" [ 1013.348655] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.357644] env[63531]: DEBUG oslo_vmware.api [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118410, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.362770] env[63531]: DEBUG nova.scheduler.client.report [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1013.560531] env[63531]: DEBUG nova.compute.manager [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: afa5f2c8-e519-4eb7-8e0e-ada9efb630e7] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1013.587241] env[63531]: DEBUG nova.virt.hardware [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1013.587497] env[63531]: DEBUG nova.virt.hardware [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1013.587661] env[63531]: DEBUG nova.virt.hardware [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1013.587849] env[63531]: DEBUG nova.virt.hardware [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1013.588037] env[63531]: DEBUG nova.virt.hardware [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1013.588213] env[63531]: DEBUG nova.virt.hardware [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1013.588432] env[63531]: DEBUG nova.virt.hardware [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1013.588602] env[63531]: DEBUG nova.virt.hardware [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1013.588774] env[63531]: DEBUG nova.virt.hardware [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1013.588967] env[63531]: DEBUG nova.virt.hardware [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1013.589203] env[63531]: DEBUG nova.virt.hardware [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1013.590084] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73e5fc71-58d9-4f05-b9d1-30f15672de8c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.598413] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57970d41-9e3a-4e36-811d-b40ebd6d1f73 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.776609] env[63531]: DEBUG oslo_vmware.api [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118408, 'name': ReconfigVM_Task, 'duration_secs': 0.549395} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.776905] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Reconfigured VM instance instance-0000005e to attach disk [datastore2] 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7/5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1013.777573] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-12919ff3-1621-44f0-aba8-c9aa78437563 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.784878] env[63531]: DEBUG oslo_vmware.api [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 1013.784878] env[63531]: value = "task-1118411" [ 1013.784878] env[63531]: _type = "Task" [ 1013.784878] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.794156] env[63531]: DEBUG oslo_vmware.api [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118411, 'name': Rename_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.818895] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Releasing lock "refresh_cache-3fe7a7a8-fc43-4560-b101-227907c91fd0" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.819230] env[63531]: DEBUG nova.compute.manager [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Instance network_info: |[{"id": "03e99378-6a13-403e-9cb4-856abf141914", "address": "fa:16:3e:3d:57:da", "network": {"id": "ea2d47d3-a502-4afd-9bc5-ee4a37017901", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1361282648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94a95667c00947ea9183a6307c569c90", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6c45fd-e930-495a-9cb7-df84eda443b1", "external-id": "nsx-vlan-transportzone-407", "segmentation_id": 407, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03e99378-6a", "ovs_interfaceid": "03e99378-6a13-403e-9cb4-856abf141914", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1013.819818] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3d:57:da', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2b6c45fd-e930-495a-9cb7-df84eda443b1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '03e99378-6a13-403e-9cb4-856abf141914', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1013.827301] env[63531]: DEBUG oslo.service.loopingcall [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1013.827519] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1013.827745] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-18327959-f196-45df-9374-43f27c8fd4db {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.847863] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1013.847863] env[63531]: value = "task-1118412" [ 1013.847863] env[63531]: _type = "Task" [ 1013.847863] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.859473] env[63531]: DEBUG oslo_vmware.api [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118410, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074981} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.862589] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1013.863207] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118412, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.863588] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0168aee8-061b-454a-ad17-021a15cbee0e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.868102] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.332s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.868658] env[63531]: DEBUG nova.compute.manager [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: ba97ce22-ea4f-430b-a4c6-08204a568436] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1013.880700] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 4.989s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.880906] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.881087] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63531) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1013.881429] env[63531]: DEBUG oslo_concurrency.lockutils [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.366s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.883464] env[63531]: INFO nova.compute.claims [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b6db77e6-4447-4b22-93d7-265ff0ada0bd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1013.898141] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Reconfiguring VM instance instance-00000055 to attach disk [datastore1] 911f23f7-e320-48f3-87a9-0239b013feff/911f23f7-e320-48f3-87a9-0239b013feff.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1013.899444] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3599a9c6-4311-4e2d-8573-531acc7fab0d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.902631] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e70db0fe-6960-4dbc-89e1-c295c060c7d2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.945962] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1c29f3c-d303-4be7-84f7-c353500d18c9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.951367] env[63531]: DEBUG oslo_vmware.api [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 1013.951367] env[63531]: value = "task-1118413" [ 1013.951367] env[63531]: _type = "Task" [ 1013.951367] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.965607] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-434936b5-d95f-41b3-a9ab-b775d66e9a28 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.972401] env[63531]: DEBUG oslo_vmware.api [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118413, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.978757] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf3fff6e-1044-44cf-b558-6dbc24d50b69 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.017852] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179611MB free_disk=169GB free_vcpus=48 pci_devices=None {{(pid=63531) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1014.018061] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.205570] env[63531]: DEBUG nova.network.neutron [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: afa5f2c8-e519-4eb7-8e0e-ada9efb630e7] Successfully updated port: b320eed5-1bc8-4774-9abd-e8cfcd640e56 {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1014.210877] env[63531]: DEBUG nova.compute.manager [req-e4986de9-f928-464f-a59d-05853f1b5df5 req-d01d0da9-f1b3-49c8-8164-1dfd2fcc907a service nova] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Received event network-changed-03e99378-6a13-403e-9cb4-856abf141914 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1014.212204] env[63531]: DEBUG nova.compute.manager [req-e4986de9-f928-464f-a59d-05853f1b5df5 req-d01d0da9-f1b3-49c8-8164-1dfd2fcc907a service nova] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Refreshing instance network info cache due to event network-changed-03e99378-6a13-403e-9cb4-856abf141914. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1014.212398] env[63531]: DEBUG oslo_concurrency.lockutils [req-e4986de9-f928-464f-a59d-05853f1b5df5 req-d01d0da9-f1b3-49c8-8164-1dfd2fcc907a service nova] Acquiring lock "refresh_cache-3fe7a7a8-fc43-4560-b101-227907c91fd0" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.212590] env[63531]: DEBUG oslo_concurrency.lockutils [req-e4986de9-f928-464f-a59d-05853f1b5df5 req-d01d0da9-f1b3-49c8-8164-1dfd2fcc907a service nova] Acquired lock "refresh_cache-3fe7a7a8-fc43-4560-b101-227907c91fd0" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.212944] env[63531]: DEBUG nova.network.neutron [req-e4986de9-f928-464f-a59d-05853f1b5df5 req-d01d0da9-f1b3-49c8-8164-1dfd2fcc907a service nova] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Refreshing network info cache for port 03e99378-6a13-403e-9cb4-856abf141914 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1014.297782] env[63531]: DEBUG oslo_vmware.api [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118411, 'name': Rename_Task, 'duration_secs': 0.238098} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.297782] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1014.297782] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e540a356-f920-47e0-bb47-bdf9b39552eb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.305054] env[63531]: DEBUG oslo_vmware.api [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 1014.305054] env[63531]: value = "task-1118414" [ 1014.305054] env[63531]: _type = "Task" [ 1014.305054] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.318651] env[63531]: DEBUG oslo_vmware.api [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118414, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.358954] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118412, 'name': CreateVM_Task, 'duration_secs': 0.39517} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.359178] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1014.359884] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.360171] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.360458] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1014.360717] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-75ff725e-6e66-4f64-aea6-7274f89bc07c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.366119] env[63531]: DEBUG oslo_vmware.api [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 1014.366119] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5221cdd5-b588-49eb-8f9a-3105188640c1" [ 1014.366119] env[63531]: _type = "Task" [ 1014.366119] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.378112] env[63531]: DEBUG oslo_vmware.api [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5221cdd5-b588-49eb-8f9a-3105188640c1, 'name': SearchDatastore_Task, 'duration_secs': 0.009863} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.378425] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1014.378666] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1014.378905] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.379077] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.379267] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1014.379526] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-af4da36c-2d93-471a-af82-14cc094b2930 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.387544] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1014.387724] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1014.388467] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6ce3d704-025f-44d4-b0bc-0f937a95df2b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.393703] env[63531]: DEBUG oslo_vmware.api [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 1014.393703] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52059f4d-a899-2bf8-ec8c-c38abbd8b769" [ 1014.393703] env[63531]: _type = "Task" [ 1014.393703] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.402119] env[63531]: DEBUG nova.compute.utils [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1014.405658] env[63531]: DEBUG oslo_vmware.api [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52059f4d-a899-2bf8-ec8c-c38abbd8b769, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.406145] env[63531]: DEBUG nova.compute.manager [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: ba97ce22-ea4f-430b-a4c6-08204a568436] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1014.406314] env[63531]: DEBUG nova.network.neutron [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: ba97ce22-ea4f-430b-a4c6-08204a568436] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1014.445327] env[63531]: DEBUG nova.policy [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4186052665ce457b99e1fee971be05b4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4ab5ee56531c4e638f8b695b9e5d082f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 1014.461944] env[63531]: DEBUG oslo_vmware.api [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118413, 'name': ReconfigVM_Task, 'duration_secs': 0.39453} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.462242] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Reconfigured VM instance instance-00000055 to attach disk [datastore1] 911f23f7-e320-48f3-87a9-0239b013feff/911f23f7-e320-48f3-87a9-0239b013feff.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1014.463644] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'size': 0, 'encrypted': False, 'disk_bus': None, 'guest_format': None, 'encryption_format': None, 'encryption_options': None, 'image_type': None, 'device_name': '/dev/sda', 'encryption_secret_uuid': None, 'boot_index': 0, 'device_type': 'disk', 'image_id': '9d550399-588c-4bd5-8436-eba9b523bd13'}], 'ephemerals': [], 'block_device_mapping': [{'mount_device': '/dev/sdb', 'attachment_id': 'bfb4f39b-4e05-4389-9664-3531fd144dd7', 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244776', 'volume_id': '3ac35683-e14e-4d02-8b0f-a5ea34354c7f', 'name': 'volume-3ac35683-e14e-4d02-8b0f-a5ea34354c7f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '911f23f7-e320-48f3-87a9-0239b013feff', 'attached_at': '', 'detached_at': '', 'volume_id': '3ac35683-e14e-4d02-8b0f-a5ea34354c7f', 'serial': '3ac35683-e14e-4d02-8b0f-a5ea34354c7f'}, 'disk_bus': None, 'guest_format': None, 'delete_on_termination': False, 'boot_index': None, 'device_type': None, 'volume_type': None}], 'swap': None} {{(pid=63531) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1014.463854] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Volume attach. Driver type: vmdk {{(pid=63531) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1014.464074] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244776', 'volume_id': '3ac35683-e14e-4d02-8b0f-a5ea34354c7f', 'name': 'volume-3ac35683-e14e-4d02-8b0f-a5ea34354c7f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '911f23f7-e320-48f3-87a9-0239b013feff', 'attached_at': '', 'detached_at': '', 'volume_id': '3ac35683-e14e-4d02-8b0f-a5ea34354c7f', 'serial': '3ac35683-e14e-4d02-8b0f-a5ea34354c7f'} {{(pid=63531) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1014.464873] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aedae020-642f-4969-bb19-fbba97c0e7fe {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.481486] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07fc52d6-ba18-4bdd-b037-92e7834fb2ef {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.507076] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Reconfiguring VM instance instance-00000055 to attach disk [datastore1] volume-3ac35683-e14e-4d02-8b0f-a5ea34354c7f/volume-3ac35683-e14e-4d02-8b0f-a5ea34354c7f.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1014.507372] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-27a92145-773f-4774-9482-b702a24cd98f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.526229] env[63531]: DEBUG oslo_vmware.api [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 1014.526229] env[63531]: value = "task-1118415" [ 1014.526229] env[63531]: _type = "Task" [ 1014.526229] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.534617] env[63531]: DEBUG oslo_vmware.api [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118415, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.716117] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Acquiring lock "refresh_cache-afa5f2c8-e519-4eb7-8e0e-ada9efb630e7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.716364] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Acquired lock "refresh_cache-afa5f2c8-e519-4eb7-8e0e-ada9efb630e7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.716493] env[63531]: DEBUG nova.network.neutron [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: afa5f2c8-e519-4eb7-8e0e-ada9efb630e7] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1014.784279] env[63531]: DEBUG nova.network.neutron [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: ba97ce22-ea4f-430b-a4c6-08204a568436] Successfully created port: 8baef0c0-0988-4a88-8808-34fde881d31d {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1014.816426] env[63531]: DEBUG oslo_vmware.api [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118414, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.904840] env[63531]: DEBUG oslo_vmware.api [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52059f4d-a899-2bf8-ec8c-c38abbd8b769, 'name': SearchDatastore_Task, 'duration_secs': 0.008155} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.905674] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fd673558-e0f4-42ff-9076-7be27d64b8a2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.908342] env[63531]: DEBUG nova.compute.manager [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: ba97ce22-ea4f-430b-a4c6-08204a568436] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1014.917510] env[63531]: DEBUG oslo_vmware.api [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 1014.917510] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52c5daa3-8b7f-df66-c8fe-f9de52520939" [ 1014.917510] env[63531]: _type = "Task" [ 1014.917510] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.926796] env[63531]: DEBUG oslo_vmware.api [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52c5daa3-8b7f-df66-c8fe-f9de52520939, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.036260] env[63531]: DEBUG oslo_vmware.api [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118415, 'name': ReconfigVM_Task, 'duration_secs': 0.332625} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.038609] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Reconfigured VM instance instance-00000055 to attach disk [datastore1] volume-3ac35683-e14e-4d02-8b0f-a5ea34354c7f/volume-3ac35683-e14e-4d02-8b0f-a5ea34354c7f.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1015.044166] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-094a3993-2c88-4443-ae24-78ecf2f680a5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.061996] env[63531]: DEBUG oslo_vmware.api [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 1015.061996] env[63531]: value = "task-1118416" [ 1015.061996] env[63531]: _type = "Task" [ 1015.061996] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.072767] env[63531]: DEBUG oslo_vmware.api [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118416, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.083789] env[63531]: DEBUG nova.network.neutron [req-e4986de9-f928-464f-a59d-05853f1b5df5 req-d01d0da9-f1b3-49c8-8164-1dfd2fcc907a service nova] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Updated VIF entry in instance network info cache for port 03e99378-6a13-403e-9cb4-856abf141914. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1015.084195] env[63531]: DEBUG nova.network.neutron [req-e4986de9-f928-464f-a59d-05853f1b5df5 req-d01d0da9-f1b3-49c8-8164-1dfd2fcc907a service nova] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Updating instance_info_cache with network_info: [{"id": "03e99378-6a13-403e-9cb4-856abf141914", "address": "fa:16:3e:3d:57:da", "network": {"id": "ea2d47d3-a502-4afd-9bc5-ee4a37017901", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1361282648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94a95667c00947ea9183a6307c569c90", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6c45fd-e930-495a-9cb7-df84eda443b1", "external-id": "nsx-vlan-transportzone-407", "segmentation_id": 407, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03e99378-6a", "ovs_interfaceid": "03e99378-6a13-403e-9cb4-856abf141914", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1015.143568] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b265d569-8617-4d44-a1c7-5a51f7ada60d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.151848] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcebbd75-02ec-40c0-963c-596092d49bb9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.184364] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51bd6c88-ef2c-4d9f-bbd2-6f0d316b15bb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.192502] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-777e0bee-cfd8-424e-aab2-d50bac0959d9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.207931] env[63531]: DEBUG nova.compute.provider_tree [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1015.269174] env[63531]: DEBUG nova.network.neutron [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: afa5f2c8-e519-4eb7-8e0e-ada9efb630e7] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1015.316590] env[63531]: DEBUG oslo_vmware.api [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118414, 'name': PowerOnVM_Task, 'duration_secs': 0.517923} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.316922] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1015.317151] env[63531]: INFO nova.compute.manager [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Took 8.11 seconds to spawn the instance on the hypervisor. [ 1015.317340] env[63531]: DEBUG nova.compute.manager [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1015.318139] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8722a898-808b-4eee-828b-251f31cd89c5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.428281] env[63531]: DEBUG oslo_vmware.api [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52c5daa3-8b7f-df66-c8fe-f9de52520939, 'name': SearchDatastore_Task, 'duration_secs': 0.010734} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.428997] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.428997] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 3fe7a7a8-fc43-4560-b101-227907c91fd0/3fe7a7a8-fc43-4560-b101-227907c91fd0.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1015.429389] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-60a338b2-3de8-4153-ac93-269ec92a8fcb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.472383] env[63531]: DEBUG oslo_vmware.api [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 1015.472383] env[63531]: value = "task-1118417" [ 1015.472383] env[63531]: _type = "Task" [ 1015.472383] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.472383] env[63531]: DEBUG oslo_vmware.api [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118417, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.551740] env[63531]: DEBUG nova.network.neutron [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: afa5f2c8-e519-4eb7-8e0e-ada9efb630e7] Updating instance_info_cache with network_info: [{"id": "b320eed5-1bc8-4774-9abd-e8cfcd640e56", "address": "fa:16:3e:d9:ea:e7", "network": {"id": "66c17d79-e8ec-4c9e-b6e3-5573d8902c4d", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-285539233-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ab5ee56531c4e638f8b695b9e5d082f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5446413d-c3b0-4cd2-a962-62240db178ac", "external-id": "nsx-vlan-transportzone-528", "segmentation_id": 528, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb320eed5-1b", "ovs_interfaceid": "b320eed5-1bc8-4774-9abd-e8cfcd640e56", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1015.573428] env[63531]: DEBUG oslo_vmware.api [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118416, 'name': ReconfigVM_Task, 'duration_secs': 0.151427} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.573779] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244776', 'volume_id': '3ac35683-e14e-4d02-8b0f-a5ea34354c7f', 'name': 'volume-3ac35683-e14e-4d02-8b0f-a5ea34354c7f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '911f23f7-e320-48f3-87a9-0239b013feff', 'attached_at': '', 'detached_at': '', 'volume_id': '3ac35683-e14e-4d02-8b0f-a5ea34354c7f', 'serial': '3ac35683-e14e-4d02-8b0f-a5ea34354c7f'} {{(pid=63531) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1015.574661] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7162b18f-e494-42b8-a8b1-ceca3f86494f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.581217] env[63531]: DEBUG oslo_vmware.api [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 1015.581217] env[63531]: value = "task-1118418" [ 1015.581217] env[63531]: _type = "Task" [ 1015.581217] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.586699] env[63531]: DEBUG oslo_concurrency.lockutils [req-e4986de9-f928-464f-a59d-05853f1b5df5 req-d01d0da9-f1b3-49c8-8164-1dfd2fcc907a service nova] Releasing lock "refresh_cache-3fe7a7a8-fc43-4560-b101-227907c91fd0" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.586970] env[63531]: DEBUG nova.compute.manager [req-e4986de9-f928-464f-a59d-05853f1b5df5 req-d01d0da9-f1b3-49c8-8164-1dfd2fcc907a service nova] [instance: afa5f2c8-e519-4eb7-8e0e-ada9efb630e7] Received event network-vif-plugged-b320eed5-1bc8-4774-9abd-e8cfcd640e56 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1015.587214] env[63531]: DEBUG oslo_concurrency.lockutils [req-e4986de9-f928-464f-a59d-05853f1b5df5 req-d01d0da9-f1b3-49c8-8164-1dfd2fcc907a service nova] Acquiring lock "afa5f2c8-e519-4eb7-8e0e-ada9efb630e7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1015.587395] env[63531]: DEBUG oslo_concurrency.lockutils [req-e4986de9-f928-464f-a59d-05853f1b5df5 req-d01d0da9-f1b3-49c8-8164-1dfd2fcc907a service nova] Lock "afa5f2c8-e519-4eb7-8e0e-ada9efb630e7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1015.587594] env[63531]: DEBUG oslo_concurrency.lockutils [req-e4986de9-f928-464f-a59d-05853f1b5df5 req-d01d0da9-f1b3-49c8-8164-1dfd2fcc907a service nova] Lock "afa5f2c8-e519-4eb7-8e0e-ada9efb630e7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.588021] env[63531]: DEBUG nova.compute.manager [req-e4986de9-f928-464f-a59d-05853f1b5df5 req-d01d0da9-f1b3-49c8-8164-1dfd2fcc907a service nova] [instance: afa5f2c8-e519-4eb7-8e0e-ada9efb630e7] No waiting events found dispatching network-vif-plugged-b320eed5-1bc8-4774-9abd-e8cfcd640e56 {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1015.588241] env[63531]: WARNING nova.compute.manager [req-e4986de9-f928-464f-a59d-05853f1b5df5 req-d01d0da9-f1b3-49c8-8164-1dfd2fcc907a service nova] [instance: afa5f2c8-e519-4eb7-8e0e-ada9efb630e7] Received unexpected event network-vif-plugged-b320eed5-1bc8-4774-9abd-e8cfcd640e56 for instance with vm_state building and task_state spawning. [ 1015.591916] env[63531]: DEBUG oslo_vmware.api [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118418, 'name': Rename_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.710975] env[63531]: DEBUG nova.scheduler.client.report [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1015.839583] env[63531]: INFO nova.compute.manager [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Took 23.30 seconds to build instance. [ 1015.923702] env[63531]: DEBUG nova.compute.manager [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: ba97ce22-ea4f-430b-a4c6-08204a568436] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1015.954422] env[63531]: DEBUG nova.virt.hardware [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1015.954689] env[63531]: DEBUG nova.virt.hardware [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1015.954853] env[63531]: DEBUG nova.virt.hardware [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1015.955055] env[63531]: DEBUG nova.virt.hardware [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1015.955214] env[63531]: DEBUG nova.virt.hardware [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1015.955372] env[63531]: DEBUG nova.virt.hardware [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1015.955586] env[63531]: DEBUG nova.virt.hardware [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1015.955778] env[63531]: DEBUG nova.virt.hardware [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1015.956061] env[63531]: DEBUG nova.virt.hardware [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1015.957301] env[63531]: DEBUG nova.virt.hardware [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1015.957301] env[63531]: DEBUG nova.virt.hardware [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1015.957395] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e482b81-85c4-4e53-9940-6446da3ff77c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.963367] env[63531]: DEBUG oslo_vmware.api [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118417, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.969364] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af5028a5-ace1-4dd8-b409-7207082e825b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.054996] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Releasing lock "refresh_cache-afa5f2c8-e519-4eb7-8e0e-ada9efb630e7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1016.055559] env[63531]: DEBUG nova.compute.manager [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: afa5f2c8-e519-4eb7-8e0e-ada9efb630e7] Instance network_info: |[{"id": "b320eed5-1bc8-4774-9abd-e8cfcd640e56", "address": "fa:16:3e:d9:ea:e7", "network": {"id": "66c17d79-e8ec-4c9e-b6e3-5573d8902c4d", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-285539233-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ab5ee56531c4e638f8b695b9e5d082f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5446413d-c3b0-4cd2-a962-62240db178ac", "external-id": "nsx-vlan-transportzone-528", "segmentation_id": 528, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb320eed5-1b", "ovs_interfaceid": "b320eed5-1bc8-4774-9abd-e8cfcd640e56", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1016.056222] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: afa5f2c8-e519-4eb7-8e0e-ada9efb630e7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d9:ea:e7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5446413d-c3b0-4cd2-a962-62240db178ac', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b320eed5-1bc8-4774-9abd-e8cfcd640e56', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1016.063459] env[63531]: DEBUG oslo.service.loopingcall [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1016.063704] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: afa5f2c8-e519-4eb7-8e0e-ada9efb630e7] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1016.063950] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cef8f6b1-1241-438b-8448-5de088ace943 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.087285] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1016.087285] env[63531]: value = "task-1118419" [ 1016.087285] env[63531]: _type = "Task" [ 1016.087285] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.102168] env[63531]: DEBUG oslo_vmware.api [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118418, 'name': Rename_Task, 'duration_secs': 0.459459} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.104771] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1016.104990] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118419, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.105204] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f06831d8-a634-486b-beaf-8c8e03f58e85 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.112301] env[63531]: DEBUG oslo_vmware.api [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 1016.112301] env[63531]: value = "task-1118420" [ 1016.112301] env[63531]: _type = "Task" [ 1016.112301] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.129561] env[63531]: DEBUG oslo_vmware.api [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118420, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.216768] env[63531]: DEBUG oslo_concurrency.lockutils [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.335s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.217327] env[63531]: DEBUG nova.compute.manager [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b6db77e6-4447-4b22-93d7-265ff0ada0bd] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1016.220105] env[63531]: DEBUG oslo_concurrency.lockutils [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.831s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.221679] env[63531]: INFO nova.compute.claims [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1016.236607] env[63531]: DEBUG nova.compute.manager [req-4409790b-0cb2-407b-8ec6-c4af3fc1e71b req-dd6d612b-8cc3-4d71-9753-c2f9d6a023a3 service nova] [instance: afa5f2c8-e519-4eb7-8e0e-ada9efb630e7] Received event network-changed-b320eed5-1bc8-4774-9abd-e8cfcd640e56 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1016.236819] env[63531]: DEBUG nova.compute.manager [req-4409790b-0cb2-407b-8ec6-c4af3fc1e71b req-dd6d612b-8cc3-4d71-9753-c2f9d6a023a3 service nova] [instance: afa5f2c8-e519-4eb7-8e0e-ada9efb630e7] Refreshing instance network info cache due to event network-changed-b320eed5-1bc8-4774-9abd-e8cfcd640e56. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1016.237059] env[63531]: DEBUG oslo_concurrency.lockutils [req-4409790b-0cb2-407b-8ec6-c4af3fc1e71b req-dd6d612b-8cc3-4d71-9753-c2f9d6a023a3 service nova] Acquiring lock "refresh_cache-afa5f2c8-e519-4eb7-8e0e-ada9efb630e7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1016.237218] env[63531]: DEBUG oslo_concurrency.lockutils [req-4409790b-0cb2-407b-8ec6-c4af3fc1e71b req-dd6d612b-8cc3-4d71-9753-c2f9d6a023a3 service nova] Acquired lock "refresh_cache-afa5f2c8-e519-4eb7-8e0e-ada9efb630e7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.237389] env[63531]: DEBUG nova.network.neutron [req-4409790b-0cb2-407b-8ec6-c4af3fc1e71b req-dd6d612b-8cc3-4d71-9753-c2f9d6a023a3 service nova] [instance: afa5f2c8-e519-4eb7-8e0e-ada9efb630e7] Refreshing network info cache for port b320eed5-1bc8-4774-9abd-e8cfcd640e56 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1016.342928] env[63531]: DEBUG oslo_concurrency.lockutils [None req-00525ba4-7a5d-4589-961e-e3d6828789b7 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.835s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.439856] env[63531]: DEBUG nova.network.neutron [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: ba97ce22-ea4f-430b-a4c6-08204a568436] Successfully updated port: 8baef0c0-0988-4a88-8808-34fde881d31d {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1016.456927] env[63531]: DEBUG oslo_vmware.api [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118417, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.525418} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.457226] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 3fe7a7a8-fc43-4560-b101-227907c91fd0/3fe7a7a8-fc43-4560-b101-227907c91fd0.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1016.457450] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1016.457718] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ee7fb971-8a3c-452f-9a02-0bd60cc64b14 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.465318] env[63531]: DEBUG oslo_vmware.api [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 1016.465318] env[63531]: value = "task-1118421" [ 1016.465318] env[63531]: _type = "Task" [ 1016.465318] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.474277] env[63531]: DEBUG oslo_vmware.api [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118421, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.598429] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118419, 'name': CreateVM_Task} progress is 25%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.623140] env[63531]: DEBUG oslo_vmware.api [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118420, 'name': PowerOnVM_Task} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.623140] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1016.623140] env[63531]: DEBUG nova.compute.manager [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1016.623582] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e367d99d-7485-4950-868c-0cc3bd733e8d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.726352] env[63531]: DEBUG nova.compute.utils [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1016.730364] env[63531]: DEBUG nova.compute.manager [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b6db77e6-4447-4b22-93d7-265ff0ada0bd] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1016.730531] env[63531]: DEBUG nova.network.neutron [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b6db77e6-4447-4b22-93d7-265ff0ada0bd] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1016.824019] env[63531]: DEBUG nova.policy [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7140e12106504b3e95ff49786361771b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '63900312a30b4942854f552758c039cf', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 1016.943228] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Acquiring lock "refresh_cache-ba97ce22-ea4f-430b-a4c6-08204a568436" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1016.943525] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Acquired lock "refresh_cache-ba97ce22-ea4f-430b-a4c6-08204a568436" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.943740] env[63531]: DEBUG nova.network.neutron [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: ba97ce22-ea4f-430b-a4c6-08204a568436] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1016.978685] env[63531]: DEBUG oslo_vmware.api [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118421, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.110026} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.979217] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1016.980925] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adeeb57b-5ae0-450a-8314-cb9bb319fd8a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.011050] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Reconfiguring VM instance instance-0000005f to attach disk [datastore1] 3fe7a7a8-fc43-4560-b101-227907c91fd0/3fe7a7a8-fc43-4560-b101-227907c91fd0.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1017.013833] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c2f1055e-ce80-486f-bda2-3b35e95de9d5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.038099] env[63531]: DEBUG oslo_vmware.api [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 1017.038099] env[63531]: value = "task-1118422" [ 1017.038099] env[63531]: _type = "Task" [ 1017.038099] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.053460] env[63531]: DEBUG oslo_vmware.api [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118422, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.102473] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118419, 'name': CreateVM_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.151444] env[63531]: DEBUG oslo_concurrency.lockutils [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.151628] env[63531]: DEBUG nova.compute.manager [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Stashing vm_state: active {{(pid=63531) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1017.197562] env[63531]: DEBUG nova.network.neutron [req-4409790b-0cb2-407b-8ec6-c4af3fc1e71b req-dd6d612b-8cc3-4d71-9753-c2f9d6a023a3 service nova] [instance: afa5f2c8-e519-4eb7-8e0e-ada9efb630e7] Updated VIF entry in instance network info cache for port b320eed5-1bc8-4774-9abd-e8cfcd640e56. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1017.198203] env[63531]: DEBUG nova.network.neutron [req-4409790b-0cb2-407b-8ec6-c4af3fc1e71b req-dd6d612b-8cc3-4d71-9753-c2f9d6a023a3 service nova] [instance: afa5f2c8-e519-4eb7-8e0e-ada9efb630e7] Updating instance_info_cache with network_info: [{"id": "b320eed5-1bc8-4774-9abd-e8cfcd640e56", "address": "fa:16:3e:d9:ea:e7", "network": {"id": "66c17d79-e8ec-4c9e-b6e3-5573d8902c4d", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-285539233-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ab5ee56531c4e638f8b695b9e5d082f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5446413d-c3b0-4cd2-a962-62240db178ac", "external-id": "nsx-vlan-transportzone-528", "segmentation_id": 528, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb320eed5-1b", "ovs_interfaceid": "b320eed5-1bc8-4774-9abd-e8cfcd640e56", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.230756] env[63531]: DEBUG nova.compute.manager [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b6db77e6-4447-4b22-93d7-265ff0ada0bd] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1017.381984] env[63531]: DEBUG nova.network.neutron [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b6db77e6-4447-4b22-93d7-265ff0ada0bd] Successfully created port: d6c4d45f-8dc2-4f95-947f-9320df584daf {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1017.497224] env[63531]: DEBUG nova.network.neutron [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: ba97ce22-ea4f-430b-a4c6-08204a568436] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1017.513094] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7f67bb2-b18e-48a0-9bb4-fd028f7351b4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.523268] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-598b7588-9138-4a82-9cd3-36c334457f60 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.562498] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17b9b2c5-7c4b-46e4-86ab-67f647e6d892 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.581594] env[63531]: DEBUG oslo_vmware.api [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118422, 'name': ReconfigVM_Task, 'duration_secs': 0.459135} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.581995] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Reconfigured VM instance instance-0000005f to attach disk [datastore1] 3fe7a7a8-fc43-4560-b101-227907c91fd0/3fe7a7a8-fc43-4560-b101-227907c91fd0.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1017.584238] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ff53d1f-038c-44aa-aca1-25b1dce5467a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.589516] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1791e081-2c14-4b1c-b948-b139e54afe31 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.603692] env[63531]: DEBUG nova.compute.provider_tree [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1017.607055] env[63531]: DEBUG oslo_vmware.api [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 1017.607055] env[63531]: value = "task-1118423" [ 1017.607055] env[63531]: _type = "Task" [ 1017.607055] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.615718] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118419, 'name': CreateVM_Task, 'duration_secs': 1.135373} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.619326] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: afa5f2c8-e519-4eb7-8e0e-ada9efb630e7] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1017.620291] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1017.620492] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.620827] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1017.621393] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5d5a95cc-3b82-497d-a232-0a8a0fc7c6a7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.626699] env[63531]: DEBUG oslo_vmware.api [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118423, 'name': Rename_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.632390] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Waiting for the task: (returnval){ [ 1017.632390] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52b4c4b3-270c-99df-2458-8579a7cff69c" [ 1017.632390] env[63531]: _type = "Task" [ 1017.632390] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.643912] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52b4c4b3-270c-99df-2458-8579a7cff69c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.680276] env[63531]: DEBUG oslo_concurrency.lockutils [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.701585] env[63531]: DEBUG oslo_concurrency.lockutils [req-4409790b-0cb2-407b-8ec6-c4af3fc1e71b req-dd6d612b-8cc3-4d71-9753-c2f9d6a023a3 service nova] Releasing lock "refresh_cache-afa5f2c8-e519-4eb7-8e0e-ada9efb630e7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1017.830571] env[63531]: DEBUG nova.network.neutron [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: ba97ce22-ea4f-430b-a4c6-08204a568436] Updating instance_info_cache with network_info: [{"id": "8baef0c0-0988-4a88-8808-34fde881d31d", "address": "fa:16:3e:67:9f:13", "network": {"id": "66c17d79-e8ec-4c9e-b6e3-5573d8902c4d", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-285539233-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ab5ee56531c4e638f8b695b9e5d082f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5446413d-c3b0-4cd2-a962-62240db178ac", "external-id": "nsx-vlan-transportzone-528", "segmentation_id": 528, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8baef0c0-09", "ovs_interfaceid": "8baef0c0-0988-4a88-8808-34fde881d31d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1018.070826] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1dd6bec8-61cf-4da8-a220-2974feaa8dfc tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquiring lock "911f23f7-e320-48f3-87a9-0239b013feff" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.071235] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1dd6bec8-61cf-4da8-a220-2974feaa8dfc tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "911f23f7-e320-48f3-87a9-0239b013feff" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.108964] env[63531]: DEBUG nova.scheduler.client.report [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1018.126613] env[63531]: DEBUG oslo_vmware.api [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118423, 'name': Rename_Task, 'duration_secs': 0.386156} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.127533] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1018.127786] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4cb24cf2-b92b-408f-b059-0f02547b585e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.138074] env[63531]: DEBUG oslo_vmware.api [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 1018.138074] env[63531]: value = "task-1118424" [ 1018.138074] env[63531]: _type = "Task" [ 1018.138074] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.144997] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52b4c4b3-270c-99df-2458-8579a7cff69c, 'name': SearchDatastore_Task, 'duration_secs': 0.024678} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.145603] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.145872] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: afa5f2c8-e519-4eb7-8e0e-ada9efb630e7] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1018.146139] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.146295] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.146481] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1018.146999] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-86a7984e-4b41-4dc2-9751-81ae779d3b06 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.151557] env[63531]: DEBUG oslo_vmware.api [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118424, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.158530] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1018.158650] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1018.159528] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a84bed98-887b-4957-810d-a143e6c6b3d1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.164987] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Waiting for the task: (returnval){ [ 1018.164987] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]521d1424-2f45-418e-94d5-6b1a1f2c9d08" [ 1018.164987] env[63531]: _type = "Task" [ 1018.164987] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.173417] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]521d1424-2f45-418e-94d5-6b1a1f2c9d08, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.242546] env[63531]: DEBUG nova.compute.manager [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b6db77e6-4447-4b22-93d7-265ff0ada0bd] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1018.271432] env[63531]: DEBUG nova.virt.hardware [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1018.272373] env[63531]: DEBUG nova.virt.hardware [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1018.272573] env[63531]: DEBUG nova.virt.hardware [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1018.272782] env[63531]: DEBUG nova.virt.hardware [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1018.272940] env[63531]: DEBUG nova.virt.hardware [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1018.273116] env[63531]: DEBUG nova.virt.hardware [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1018.273343] env[63531]: DEBUG nova.virt.hardware [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1018.273515] env[63531]: DEBUG nova.virt.hardware [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1018.273693] env[63531]: DEBUG nova.virt.hardware [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1018.273895] env[63531]: DEBUG nova.virt.hardware [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1018.274116] env[63531]: DEBUG nova.virt.hardware [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1018.274990] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a22c2f16-988c-439a-9afa-8bc6d299fd0d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.280357] env[63531]: DEBUG nova.compute.manager [req-237cdb70-f304-4a5b-9364-30aed966e80d req-3177388c-dde7-46c2-b20f-4c85c526c55d service nova] [instance: ba97ce22-ea4f-430b-a4c6-08204a568436] Received event network-vif-plugged-8baef0c0-0988-4a88-8808-34fde881d31d {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1018.280583] env[63531]: DEBUG oslo_concurrency.lockutils [req-237cdb70-f304-4a5b-9364-30aed966e80d req-3177388c-dde7-46c2-b20f-4c85c526c55d service nova] Acquiring lock "ba97ce22-ea4f-430b-a4c6-08204a568436-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.280799] env[63531]: DEBUG oslo_concurrency.lockutils [req-237cdb70-f304-4a5b-9364-30aed966e80d req-3177388c-dde7-46c2-b20f-4c85c526c55d service nova] Lock "ba97ce22-ea4f-430b-a4c6-08204a568436-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.280971] env[63531]: DEBUG oslo_concurrency.lockutils [req-237cdb70-f304-4a5b-9364-30aed966e80d req-3177388c-dde7-46c2-b20f-4c85c526c55d service nova] Lock "ba97ce22-ea4f-430b-a4c6-08204a568436-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.281223] env[63531]: DEBUG nova.compute.manager [req-237cdb70-f304-4a5b-9364-30aed966e80d req-3177388c-dde7-46c2-b20f-4c85c526c55d service nova] [instance: ba97ce22-ea4f-430b-a4c6-08204a568436] No waiting events found dispatching network-vif-plugged-8baef0c0-0988-4a88-8808-34fde881d31d {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1018.281365] env[63531]: WARNING nova.compute.manager [req-237cdb70-f304-4a5b-9364-30aed966e80d req-3177388c-dde7-46c2-b20f-4c85c526c55d service nova] [instance: ba97ce22-ea4f-430b-a4c6-08204a568436] Received unexpected event network-vif-plugged-8baef0c0-0988-4a88-8808-34fde881d31d for instance with vm_state building and task_state spawning. [ 1018.281533] env[63531]: DEBUG nova.compute.manager [req-237cdb70-f304-4a5b-9364-30aed966e80d req-3177388c-dde7-46c2-b20f-4c85c526c55d service nova] [instance: ba97ce22-ea4f-430b-a4c6-08204a568436] Received event network-changed-8baef0c0-0988-4a88-8808-34fde881d31d {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1018.281726] env[63531]: DEBUG nova.compute.manager [req-237cdb70-f304-4a5b-9364-30aed966e80d req-3177388c-dde7-46c2-b20f-4c85c526c55d service nova] [instance: ba97ce22-ea4f-430b-a4c6-08204a568436] Refreshing instance network info cache due to event network-changed-8baef0c0-0988-4a88-8808-34fde881d31d. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1018.281908] env[63531]: DEBUG oslo_concurrency.lockutils [req-237cdb70-f304-4a5b-9364-30aed966e80d req-3177388c-dde7-46c2-b20f-4c85c526c55d service nova] Acquiring lock "refresh_cache-ba97ce22-ea4f-430b-a4c6-08204a568436" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.288826] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e87f7405-f29b-45b8-ac08-cd357396c83a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.333208] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Releasing lock "refresh_cache-ba97ce22-ea4f-430b-a4c6-08204a568436" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.333554] env[63531]: DEBUG nova.compute.manager [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: ba97ce22-ea4f-430b-a4c6-08204a568436] Instance network_info: |[{"id": "8baef0c0-0988-4a88-8808-34fde881d31d", "address": "fa:16:3e:67:9f:13", "network": {"id": "66c17d79-e8ec-4c9e-b6e3-5573d8902c4d", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-285539233-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ab5ee56531c4e638f8b695b9e5d082f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5446413d-c3b0-4cd2-a962-62240db178ac", "external-id": "nsx-vlan-transportzone-528", "segmentation_id": 528, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8baef0c0-09", "ovs_interfaceid": "8baef0c0-0988-4a88-8808-34fde881d31d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1018.333906] env[63531]: DEBUG oslo_concurrency.lockutils [req-237cdb70-f304-4a5b-9364-30aed966e80d req-3177388c-dde7-46c2-b20f-4c85c526c55d service nova] Acquired lock "refresh_cache-ba97ce22-ea4f-430b-a4c6-08204a568436" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.334122] env[63531]: DEBUG nova.network.neutron [req-237cdb70-f304-4a5b-9364-30aed966e80d req-3177388c-dde7-46c2-b20f-4c85c526c55d service nova] [instance: ba97ce22-ea4f-430b-a4c6-08204a568436] Refreshing network info cache for port 8baef0c0-0988-4a88-8808-34fde881d31d {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1018.335412] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: ba97ce22-ea4f-430b-a4c6-08204a568436] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:67:9f:13', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5446413d-c3b0-4cd2-a962-62240db178ac', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8baef0c0-0988-4a88-8808-34fde881d31d', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1018.342956] env[63531]: DEBUG oslo.service.loopingcall [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1018.344028] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ba97ce22-ea4f-430b-a4c6-08204a568436] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1018.344132] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7e7eab35-96d0-4c94-a742-38048b70c464 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.365077] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1018.365077] env[63531]: value = "task-1118425" [ 1018.365077] env[63531]: _type = "Task" [ 1018.365077] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.374146] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118425, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.574992] env[63531]: INFO nova.compute.manager [None req-1dd6bec8-61cf-4da8-a220-2974feaa8dfc tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Detaching volume 3ac35683-e14e-4d02-8b0f-a5ea34354c7f [ 1018.612865] env[63531]: INFO nova.virt.block_device [None req-1dd6bec8-61cf-4da8-a220-2974feaa8dfc tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Attempting to driver detach volume 3ac35683-e14e-4d02-8b0f-a5ea34354c7f from mountpoint /dev/sdb [ 1018.613152] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-1dd6bec8-61cf-4da8-a220-2974feaa8dfc tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Volume detach. Driver type: vmdk {{(pid=63531) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1018.613355] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-1dd6bec8-61cf-4da8-a220-2974feaa8dfc tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244776', 'volume_id': '3ac35683-e14e-4d02-8b0f-a5ea34354c7f', 'name': 'volume-3ac35683-e14e-4d02-8b0f-a5ea34354c7f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '911f23f7-e320-48f3-87a9-0239b013feff', 'attached_at': '', 'detached_at': '', 'volume_id': '3ac35683-e14e-4d02-8b0f-a5ea34354c7f', 'serial': '3ac35683-e14e-4d02-8b0f-a5ea34354c7f'} {{(pid=63531) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1018.614267] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3697a3cd-b432-4390-b144-f4606e9484cd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.618231] env[63531]: DEBUG oslo_concurrency.lockutils [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.398s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.618821] env[63531]: DEBUG nova.compute.manager [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1018.622029] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 4.604s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.646869] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dad11f8f-7f9d-422b-b839-cb61e7511a6f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.655612] env[63531]: DEBUG oslo_vmware.api [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118424, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.658316] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77f1f699-a32a-4cac-96b1-0c6e697fe616 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.685721] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eb4ad2f-a7b4-455b-a43f-e822dee56f72 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.696381] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]521d1424-2f45-418e-94d5-6b1a1f2c9d08, 'name': SearchDatastore_Task, 'duration_secs': 0.011151} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.707563] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-1dd6bec8-61cf-4da8-a220-2974feaa8dfc tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] The volume has not been displaced from its original location: [datastore1] volume-3ac35683-e14e-4d02-8b0f-a5ea34354c7f/volume-3ac35683-e14e-4d02-8b0f-a5ea34354c7f.vmdk. No consolidation needed. {{(pid=63531) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1018.712783] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-1dd6bec8-61cf-4da8-a220-2974feaa8dfc tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Reconfiguring VM instance instance-00000055 to detach disk 2001 {{(pid=63531) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1018.713168] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a70e1e37-c087-469e-b985-510c18e6019b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.715528] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bd89ecf1-87d1-4f88-b23b-49f3e418affe {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.732848] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Waiting for the task: (returnval){ [ 1018.732848] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52f21c98-d8c8-38c8-bab4-da9fb5fdbf44" [ 1018.732848] env[63531]: _type = "Task" [ 1018.732848] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.737376] env[63531]: DEBUG oslo_vmware.api [None req-1dd6bec8-61cf-4da8-a220-2974feaa8dfc tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 1018.737376] env[63531]: value = "task-1118426" [ 1018.737376] env[63531]: _type = "Task" [ 1018.737376] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.743942] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52f21c98-d8c8-38c8-bab4-da9fb5fdbf44, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.748947] env[63531]: DEBUG oslo_vmware.api [None req-1dd6bec8-61cf-4da8-a220-2974feaa8dfc tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118426, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.879610] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118425, 'name': CreateVM_Task, 'duration_secs': 0.424927} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.879814] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ba97ce22-ea4f-430b-a4c6-08204a568436] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1018.880508] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.880675] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.881011] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1018.881279] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7bb8cf58-208d-410e-8ba6-e486144a25a2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.887472] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Waiting for the task: (returnval){ [ 1018.887472] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5245922f-b30f-667f-2690-b5b261f2fd3c" [ 1018.887472] env[63531]: _type = "Task" [ 1018.887472] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.895925] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5245922f-b30f-667f-2690-b5b261f2fd3c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.125232] env[63531]: DEBUG nova.compute.utils [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1019.135288] env[63531]: DEBUG nova.compute.manager [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1019.135726] env[63531]: DEBUG nova.network.neutron [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1019.152609] env[63531]: DEBUG oslo_vmware.api [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118424, 'name': PowerOnVM_Task, 'duration_secs': 0.752194} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.152898] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1019.153123] env[63531]: INFO nova.compute.manager [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Took 7.89 seconds to spawn the instance on the hypervisor. [ 1019.153311] env[63531]: DEBUG nova.compute.manager [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1019.154141] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74669540-18c3-474b-8696-36738c6c8b50 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.185781] env[63531]: DEBUG nova.policy [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '597d424341124b7db23dc7a104107148', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '308eabafea9c4188a58a0f1c22074d2f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 1019.246056] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52f21c98-d8c8-38c8-bab4-da9fb5fdbf44, 'name': SearchDatastore_Task, 'duration_secs': 0.011839} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.246676] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.247042] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] afa5f2c8-e519-4eb7-8e0e-ada9efb630e7/afa5f2c8-e519-4eb7-8e0e-ada9efb630e7.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1019.247717] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ff1b1b74-cf30-489e-be4c-1f2ab792265e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.252103] env[63531]: DEBUG oslo_vmware.api [None req-1dd6bec8-61cf-4da8-a220-2974feaa8dfc tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118426, 'name': ReconfigVM_Task, 'duration_secs': 0.307973} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.252678] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-1dd6bec8-61cf-4da8-a220-2974feaa8dfc tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Reconfigured VM instance instance-00000055 to detach disk 2001 {{(pid=63531) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1019.258215] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a12a4f00-f859-4b07-83cb-e37a1fbb3c98 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.268159] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Waiting for the task: (returnval){ [ 1019.268159] env[63531]: value = "task-1118427" [ 1019.268159] env[63531]: _type = "Task" [ 1019.268159] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.273943] env[63531]: DEBUG oslo_vmware.api [None req-1dd6bec8-61cf-4da8-a220-2974feaa8dfc tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 1019.273943] env[63531]: value = "task-1118428" [ 1019.273943] env[63531]: _type = "Task" [ 1019.273943] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.277304] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118427, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.285965] env[63531]: DEBUG oslo_vmware.api [None req-1dd6bec8-61cf-4da8-a220-2974feaa8dfc tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118428, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.341380] env[63531]: DEBUG nova.network.neutron [req-237cdb70-f304-4a5b-9364-30aed966e80d req-3177388c-dde7-46c2-b20f-4c85c526c55d service nova] [instance: ba97ce22-ea4f-430b-a4c6-08204a568436] Updated VIF entry in instance network info cache for port 8baef0c0-0988-4a88-8808-34fde881d31d. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1019.341709] env[63531]: DEBUG nova.network.neutron [req-237cdb70-f304-4a5b-9364-30aed966e80d req-3177388c-dde7-46c2-b20f-4c85c526c55d service nova] [instance: ba97ce22-ea4f-430b-a4c6-08204a568436] Updating instance_info_cache with network_info: [{"id": "8baef0c0-0988-4a88-8808-34fde881d31d", "address": "fa:16:3e:67:9f:13", "network": {"id": "66c17d79-e8ec-4c9e-b6e3-5573d8902c4d", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-285539233-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ab5ee56531c4e638f8b695b9e5d082f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5446413d-c3b0-4cd2-a962-62240db178ac", "external-id": "nsx-vlan-transportzone-528", "segmentation_id": 528, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8baef0c0-09", "ovs_interfaceid": "8baef0c0-0988-4a88-8808-34fde881d31d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1019.398491] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5245922f-b30f-667f-2690-b5b261f2fd3c, 'name': SearchDatastore_Task, 'duration_secs': 0.017917} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.398953] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.399268] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: ba97ce22-ea4f-430b-a4c6-08204a568436] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1019.399550] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1019.399740] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1019.400049] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1019.400374] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-32cfd5fb-62a5-4f9a-be88-858b2076f40e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.410137] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1019.410137] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1019.410820] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-72b25e3d-0943-4684-b3a5-d552b4b40ebe {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.416730] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Waiting for the task: (returnval){ [ 1019.416730] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52fa6db9-52d0-a993-8eaf-1da67bc2663d" [ 1019.416730] env[63531]: _type = "Task" [ 1019.416730] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.426965] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52fa6db9-52d0-a993-8eaf-1da67bc2663d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.478911] env[63531]: DEBUG nova.network.neutron [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b6db77e6-4447-4b22-93d7-265ff0ada0bd] Successfully updated port: d6c4d45f-8dc2-4f95-947f-9320df584daf {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1019.540684] env[63531]: DEBUG nova.network.neutron [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Successfully created port: b731766f-738c-4d77-a00e-98ed6c683e55 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1019.636896] env[63531]: DEBUG nova.compute.manager [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1019.642120] env[63531]: INFO nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Updating resource usage from migration 2e83b091-f95e-45ca-9edc-3ab1f0e9a94a [ 1019.680929] env[63531]: INFO nova.compute.manager [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Took 20.63 seconds to build instance. [ 1019.686029] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 8dec0cac-defc-42f2-ab56-6b3ae60ad858 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1019.686029] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1019.686029] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 879924d4-b465-4102-a0e3-c7b2be7ef08b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1019.686029] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1019.686029] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance ccf00b09-29fe-4a76-a8af-97774acd77de actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1019.686029] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 911f23f7-e320-48f3-87a9-0239b013feff actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1019.686029] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 4757625a-8398-4b2d-b6fe-40eea8913068 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1019.686029] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance e742d3af-5a45-4274-adfb-c6138bf84d6a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1019.686029] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 18057005-1ff1-4e3d-bccc-8c89755d6db7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1019.686029] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance b7b2b7ac-2321-4cf3-b2ed-4877b893af63 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1019.686029] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 3fe7a7a8-fc43-4560-b101-227907c91fd0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1019.686029] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance afa5f2c8-e519-4eb7-8e0e-ada9efb630e7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1019.686029] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance ba97ce22-ea4f-430b-a4c6-08204a568436 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1019.686029] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance b6db77e6-4447-4b22-93d7-265ff0ada0bd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1019.687254] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 96c2910b-1e25-4053-9ad9-c155402ef0ac actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1019.687591] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Migration 2e83b091-f95e-45ca-9edc-3ab1f0e9a94a is active on this compute host and has allocations in placement: {'resources': {'VCPU': 1, 'MEMORY_MB': 192, 'DISK_GB': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1019.687892] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1019.688354] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Total usable vcpus: 48, total allocated vcpus: 16 {{(pid=63531) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1019.688628] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3584MB phys_disk=200GB used_disk=16GB total_vcpus=48 used_vcpus=16 pci_stats=[] {{(pid=63531) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1019.779835] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118427, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.491077} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.785064] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] afa5f2c8-e519-4eb7-8e0e-ada9efb630e7/afa5f2c8-e519-4eb7-8e0e-ada9efb630e7.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1019.785469] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: afa5f2c8-e519-4eb7-8e0e-ada9efb630e7] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1019.785855] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-eec6ef26-35d9-4e43-84c1-88a97a24b641 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.793638] env[63531]: DEBUG oslo_vmware.api [None req-1dd6bec8-61cf-4da8-a220-2974feaa8dfc tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118428, 'name': ReconfigVM_Task, 'duration_secs': 0.167924} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.795024] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-1dd6bec8-61cf-4da8-a220-2974feaa8dfc tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244776', 'volume_id': '3ac35683-e14e-4d02-8b0f-a5ea34354c7f', 'name': 'volume-3ac35683-e14e-4d02-8b0f-a5ea34354c7f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '911f23f7-e320-48f3-87a9-0239b013feff', 'attached_at': '', 'detached_at': '', 'volume_id': '3ac35683-e14e-4d02-8b0f-a5ea34354c7f', 'serial': '3ac35683-e14e-4d02-8b0f-a5ea34354c7f'} {{(pid=63531) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1019.797483] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Waiting for the task: (returnval){ [ 1019.797483] env[63531]: value = "task-1118429" [ 1019.797483] env[63531]: _type = "Task" [ 1019.797483] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.812383] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118429, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.845621] env[63531]: DEBUG oslo_concurrency.lockutils [req-237cdb70-f304-4a5b-9364-30aed966e80d req-3177388c-dde7-46c2-b20f-4c85c526c55d service nova] Releasing lock "refresh_cache-ba97ce22-ea4f-430b-a4c6-08204a568436" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.929094] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52fa6db9-52d0-a993-8eaf-1da67bc2663d, 'name': SearchDatastore_Task, 'duration_secs': 0.009884} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.933810] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4869a768-5b3a-47af-be99-39517b7bd381 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.939976] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Waiting for the task: (returnval){ [ 1019.939976] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52e7f3c8-45b4-818d-2422-aa37ed64f99e" [ 1019.939976] env[63531]: _type = "Task" [ 1019.939976] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.944814] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8a38cce-7926-4567-875d-b0c1e4df33e4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.952293] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52e7f3c8-45b4-818d-2422-aa37ed64f99e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.955059] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7bcdf87-60ca-4383-b049-79271ec2a9bf {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.987511] env[63531]: DEBUG oslo_concurrency.lockutils [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "refresh_cache-b6db77e6-4447-4b22-93d7-265ff0ada0bd" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1019.987699] env[63531]: DEBUG oslo_concurrency.lockutils [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquired lock "refresh_cache-b6db77e6-4447-4b22-93d7-265ff0ada0bd" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1019.987861] env[63531]: DEBUG nova.network.neutron [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b6db77e6-4447-4b22-93d7-265ff0ada0bd] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1019.989976] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa152f12-f343-4fb5-9ac3-108e275eb249 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.000538] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51750d67-82c7-4682-820a-b9ec290b740e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.014623] env[63531]: DEBUG nova.compute.provider_tree [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1020.183195] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8776afe7-0cc5-41f9-83c5-cc4a78d0071a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "3fe7a7a8-fc43-4560-b101-227907c91fd0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.139s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.294210] env[63531]: DEBUG nova.compute.manager [req-2f3bb8d7-68dc-49a3-af62-c584e982564b req-c58376c0-e5a3-4bcd-804e-167058fb4a19 service nova] [instance: b6db77e6-4447-4b22-93d7-265ff0ada0bd] Received event network-vif-plugged-d6c4d45f-8dc2-4f95-947f-9320df584daf {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1020.294422] env[63531]: DEBUG oslo_concurrency.lockutils [req-2f3bb8d7-68dc-49a3-af62-c584e982564b req-c58376c0-e5a3-4bcd-804e-167058fb4a19 service nova] Acquiring lock "b6db77e6-4447-4b22-93d7-265ff0ada0bd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.294631] env[63531]: DEBUG oslo_concurrency.lockutils [req-2f3bb8d7-68dc-49a3-af62-c584e982564b req-c58376c0-e5a3-4bcd-804e-167058fb4a19 service nova] Lock "b6db77e6-4447-4b22-93d7-265ff0ada0bd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.294802] env[63531]: DEBUG oslo_concurrency.lockutils [req-2f3bb8d7-68dc-49a3-af62-c584e982564b req-c58376c0-e5a3-4bcd-804e-167058fb4a19 service nova] Lock "b6db77e6-4447-4b22-93d7-265ff0ada0bd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.294968] env[63531]: DEBUG nova.compute.manager [req-2f3bb8d7-68dc-49a3-af62-c584e982564b req-c58376c0-e5a3-4bcd-804e-167058fb4a19 service nova] [instance: b6db77e6-4447-4b22-93d7-265ff0ada0bd] No waiting events found dispatching network-vif-plugged-d6c4d45f-8dc2-4f95-947f-9320df584daf {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1020.295151] env[63531]: WARNING nova.compute.manager [req-2f3bb8d7-68dc-49a3-af62-c584e982564b req-c58376c0-e5a3-4bcd-804e-167058fb4a19 service nova] [instance: b6db77e6-4447-4b22-93d7-265ff0ada0bd] Received unexpected event network-vif-plugged-d6c4d45f-8dc2-4f95-947f-9320df584daf for instance with vm_state building and task_state spawning. [ 1020.295316] env[63531]: DEBUG nova.compute.manager [req-2f3bb8d7-68dc-49a3-af62-c584e982564b req-c58376c0-e5a3-4bcd-804e-167058fb4a19 service nova] [instance: b6db77e6-4447-4b22-93d7-265ff0ada0bd] Received event network-changed-d6c4d45f-8dc2-4f95-947f-9320df584daf {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1020.295473] env[63531]: DEBUG nova.compute.manager [req-2f3bb8d7-68dc-49a3-af62-c584e982564b req-c58376c0-e5a3-4bcd-804e-167058fb4a19 service nova] [instance: b6db77e6-4447-4b22-93d7-265ff0ada0bd] Refreshing instance network info cache due to event network-changed-d6c4d45f-8dc2-4f95-947f-9320df584daf. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1020.295639] env[63531]: DEBUG oslo_concurrency.lockutils [req-2f3bb8d7-68dc-49a3-af62-c584e982564b req-c58376c0-e5a3-4bcd-804e-167058fb4a19 service nova] Acquiring lock "refresh_cache-b6db77e6-4447-4b22-93d7-265ff0ada0bd" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1020.312053] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118429, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.111955} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.312318] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: afa5f2c8-e519-4eb7-8e0e-ada9efb630e7] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1020.314024] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59e63d3d-5c4e-4958-b948-cdf180abeb78 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.335882] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: afa5f2c8-e519-4eb7-8e0e-ada9efb630e7] Reconfiguring VM instance instance-00000060 to attach disk [datastore2] afa5f2c8-e519-4eb7-8e0e-ada9efb630e7/afa5f2c8-e519-4eb7-8e0e-ada9efb630e7.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1020.336525] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f80c088a-8e70-4f6f-861e-d68431d4f7e1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.351597] env[63531]: DEBUG nova.objects.instance [None req-1dd6bec8-61cf-4da8-a220-2974feaa8dfc tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lazy-loading 'flavor' on Instance uuid 911f23f7-e320-48f3-87a9-0239b013feff {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1020.359574] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Waiting for the task: (returnval){ [ 1020.359574] env[63531]: value = "task-1118430" [ 1020.359574] env[63531]: _type = "Task" [ 1020.359574] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.368231] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "3fe7a7a8-fc43-4560-b101-227907c91fd0" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.368482] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "3fe7a7a8-fc43-4560-b101-227907c91fd0" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.368660] env[63531]: INFO nova.compute.manager [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Shelving [ 1020.369944] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118430, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.450745] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52e7f3c8-45b4-818d-2422-aa37ed64f99e, 'name': SearchDatastore_Task, 'duration_secs': 0.023881} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.451018] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.451290] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] ba97ce22-ea4f-430b-a4c6-08204a568436/ba97ce22-ea4f-430b-a4c6-08204a568436.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1020.451541] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-39b46dd1-5530-4846-8980-640e94ca10cb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.458408] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Waiting for the task: (returnval){ [ 1020.458408] env[63531]: value = "task-1118431" [ 1020.458408] env[63531]: _type = "Task" [ 1020.458408] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.465794] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118431, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.517386] env[63531]: DEBUG nova.scheduler.client.report [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1020.522513] env[63531]: DEBUG nova.network.neutron [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b6db77e6-4447-4b22-93d7-265ff0ada0bd] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1020.653831] env[63531]: DEBUG nova.compute.manager [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1020.679665] env[63531]: DEBUG nova.virt.hardware [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1020.679915] env[63531]: DEBUG nova.virt.hardware [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1020.680123] env[63531]: DEBUG nova.virt.hardware [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1020.680355] env[63531]: DEBUG nova.virt.hardware [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1020.680512] env[63531]: DEBUG nova.virt.hardware [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1020.680665] env[63531]: DEBUG nova.virt.hardware [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1020.680876] env[63531]: DEBUG nova.virt.hardware [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1020.681068] env[63531]: DEBUG nova.virt.hardware [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1020.681274] env[63531]: DEBUG nova.virt.hardware [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1020.681449] env[63531]: DEBUG nova.virt.hardware [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1020.681630] env[63531]: DEBUG nova.virt.hardware [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1020.682525] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01626815-babb-4fd9-948e-25f056b784a5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.691172] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f847d44-e3bb-412f-9ba7-55a8d38edda6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.710135] env[63531]: DEBUG nova.network.neutron [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b6db77e6-4447-4b22-93d7-265ff0ada0bd] Updating instance_info_cache with network_info: [{"id": "d6c4d45f-8dc2-4f95-947f-9320df584daf", "address": "fa:16:3e:09:83:26", "network": {"id": "80641cc0-680b-40ff-8d1b-f5a5f736a88f", "bridge": "br-int", "label": "tempest-ServersTestJSON-666288463-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "63900312a30b4942854f552758c039cf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "26472e27-9835-4f87-ab7f-ca24dfee4e83", "external-id": "nsx-vlan-transportzone-335", "segmentation_id": 335, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd6c4d45f-8d", "ovs_interfaceid": "d6c4d45f-8dc2-4f95-947f-9320df584daf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1020.874957] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118430, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.878230] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1020.878493] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bf617301-eef1-4e67-aa93-47769f9d9b46 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.885840] env[63531]: DEBUG oslo_vmware.api [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 1020.885840] env[63531]: value = "task-1118432" [ 1020.885840] env[63531]: _type = "Task" [ 1020.885840] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.895036] env[63531]: DEBUG oslo_vmware.api [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118432, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.968354] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118431, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.977174] env[63531]: DEBUG nova.compute.manager [req-672ceb9f-ad0f-4312-a869-e53703f18097 req-1308df78-1277-40c1-9be0-202ec241e136 service nova] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Received event network-vif-plugged-b731766f-738c-4d77-a00e-98ed6c683e55 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1020.977407] env[63531]: DEBUG oslo_concurrency.lockutils [req-672ceb9f-ad0f-4312-a869-e53703f18097 req-1308df78-1277-40c1-9be0-202ec241e136 service nova] Acquiring lock "96c2910b-1e25-4053-9ad9-c155402ef0ac-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.977624] env[63531]: DEBUG oslo_concurrency.lockutils [req-672ceb9f-ad0f-4312-a869-e53703f18097 req-1308df78-1277-40c1-9be0-202ec241e136 service nova] Lock "96c2910b-1e25-4053-9ad9-c155402ef0ac-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.977798] env[63531]: DEBUG oslo_concurrency.lockutils [req-672ceb9f-ad0f-4312-a869-e53703f18097 req-1308df78-1277-40c1-9be0-202ec241e136 service nova] Lock "96c2910b-1e25-4053-9ad9-c155402ef0ac-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.978013] env[63531]: DEBUG nova.compute.manager [req-672ceb9f-ad0f-4312-a869-e53703f18097 req-1308df78-1277-40c1-9be0-202ec241e136 service nova] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] No waiting events found dispatching network-vif-plugged-b731766f-738c-4d77-a00e-98ed6c683e55 {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1020.978404] env[63531]: WARNING nova.compute.manager [req-672ceb9f-ad0f-4312-a869-e53703f18097 req-1308df78-1277-40c1-9be0-202ec241e136 service nova] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Received unexpected event network-vif-plugged-b731766f-738c-4d77-a00e-98ed6c683e55 for instance with vm_state building and task_state spawning. [ 1021.022440] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63531) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1021.022557] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.401s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.022838] env[63531]: DEBUG oslo_concurrency.lockutils [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 3.872s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.023050] env[63531]: DEBUG nova.objects.instance [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63531) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1021.066713] env[63531]: DEBUG nova.network.neutron [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Successfully updated port: b731766f-738c-4d77-a00e-98ed6c683e55 {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1021.212624] env[63531]: DEBUG oslo_concurrency.lockutils [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Releasing lock "refresh_cache-b6db77e6-4447-4b22-93d7-265ff0ada0bd" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1021.213038] env[63531]: DEBUG nova.compute.manager [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b6db77e6-4447-4b22-93d7-265ff0ada0bd] Instance network_info: |[{"id": "d6c4d45f-8dc2-4f95-947f-9320df584daf", "address": "fa:16:3e:09:83:26", "network": {"id": "80641cc0-680b-40ff-8d1b-f5a5f736a88f", "bridge": "br-int", "label": "tempest-ServersTestJSON-666288463-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "63900312a30b4942854f552758c039cf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "26472e27-9835-4f87-ab7f-ca24dfee4e83", "external-id": "nsx-vlan-transportzone-335", "segmentation_id": 335, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd6c4d45f-8d", "ovs_interfaceid": "d6c4d45f-8dc2-4f95-947f-9320df584daf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1021.213285] env[63531]: DEBUG oslo_concurrency.lockutils [req-2f3bb8d7-68dc-49a3-af62-c584e982564b req-c58376c0-e5a3-4bcd-804e-167058fb4a19 service nova] Acquired lock "refresh_cache-b6db77e6-4447-4b22-93d7-265ff0ada0bd" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.213473] env[63531]: DEBUG nova.network.neutron [req-2f3bb8d7-68dc-49a3-af62-c584e982564b req-c58376c0-e5a3-4bcd-804e-167058fb4a19 service nova] [instance: b6db77e6-4447-4b22-93d7-265ff0ada0bd] Refreshing network info cache for port d6c4d45f-8dc2-4f95-947f-9320df584daf {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1021.215050] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b6db77e6-4447-4b22-93d7-265ff0ada0bd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:09:83:26', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '26472e27-9835-4f87-ab7f-ca24dfee4e83', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd6c4d45f-8dc2-4f95-947f-9320df584daf', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1021.222439] env[63531]: DEBUG oslo.service.loopingcall [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1021.223408] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b6db77e6-4447-4b22-93d7-265ff0ada0bd] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1021.223647] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ff6787a9-3402-4afd-91c0-8cca234fe4f1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.245890] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1021.245890] env[63531]: value = "task-1118433" [ 1021.245890] env[63531]: _type = "Task" [ 1021.245890] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.258152] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118433, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.360102] env[63531]: DEBUG oslo_concurrency.lockutils [None req-1dd6bec8-61cf-4da8-a220-2974feaa8dfc tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "911f23f7-e320-48f3-87a9-0239b013feff" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.289s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.371612] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118430, 'name': ReconfigVM_Task, 'duration_secs': 0.932736} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.372207] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: afa5f2c8-e519-4eb7-8e0e-ada9efb630e7] Reconfigured VM instance instance-00000060 to attach disk [datastore2] afa5f2c8-e519-4eb7-8e0e-ada9efb630e7/afa5f2c8-e519-4eb7-8e0e-ada9efb630e7.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1021.372859] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-71adbfbc-534f-4182-864c-5573a470644e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.381009] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Waiting for the task: (returnval){ [ 1021.381009] env[63531]: value = "task-1118434" [ 1021.381009] env[63531]: _type = "Task" [ 1021.381009] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.392727] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118434, 'name': Rename_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.400440] env[63531]: DEBUG oslo_vmware.api [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118432, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.472844] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118431, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.569756] env[63531]: DEBUG oslo_concurrency.lockutils [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "refresh_cache-96c2910b-1e25-4053-9ad9-c155402ef0ac" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1021.569937] env[63531]: DEBUG oslo_concurrency.lockutils [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquired lock "refresh_cache-96c2910b-1e25-4053-9ad9-c155402ef0ac" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.570128] env[63531]: DEBUG nova.network.neutron [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1021.758381] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118433, 'name': CreateVM_Task, 'duration_secs': 0.324285} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.761035] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b6db77e6-4447-4b22-93d7-265ff0ada0bd] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1021.761848] env[63531]: DEBUG oslo_concurrency.lockutils [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1021.762036] env[63531]: DEBUG oslo_concurrency.lockutils [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.762388] env[63531]: DEBUG oslo_concurrency.lockutils [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1021.763013] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0e0e2775-cfd9-4405-b623-0fcd4f7dc8cd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.769973] env[63531]: DEBUG oslo_vmware.api [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 1021.769973] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]520c063c-30a6-e5f0-0ad4-25e935549c81" [ 1021.769973] env[63531]: _type = "Task" [ 1021.769973] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.780836] env[63531]: DEBUG oslo_vmware.api [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]520c063c-30a6-e5f0-0ad4-25e935549c81, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.903024] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118434, 'name': Rename_Task, 'duration_secs': 0.172152} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.903024] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: afa5f2c8-e519-4eb7-8e0e-ada9efb630e7] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1021.903781] env[63531]: DEBUG oslo_vmware.api [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118432, 'name': PowerOffVM_Task, 'duration_secs': 0.927041} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.904159] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e0138988-78d1-4268-b7e5-fe50a4111749 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.906219] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1021.907125] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d49645b2-1128-4dad-926a-0a208169c070 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.948071] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Waiting for the task: (returnval){ [ 1021.948071] env[63531]: value = "task-1118435" [ 1021.948071] env[63531]: _type = "Task" [ 1021.948071] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.949170] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5db2c542-1e97-4d5e-89c1-770ed65981de {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.976203] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118435, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.977298] env[63531]: DEBUG nova.network.neutron [req-2f3bb8d7-68dc-49a3-af62-c584e982564b req-c58376c0-e5a3-4bcd-804e-167058fb4a19 service nova] [instance: b6db77e6-4447-4b22-93d7-265ff0ada0bd] Updated VIF entry in instance network info cache for port d6c4d45f-8dc2-4f95-947f-9320df584daf. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1021.977786] env[63531]: DEBUG nova.network.neutron [req-2f3bb8d7-68dc-49a3-af62-c584e982564b req-c58376c0-e5a3-4bcd-804e-167058fb4a19 service nova] [instance: b6db77e6-4447-4b22-93d7-265ff0ada0bd] Updating instance_info_cache with network_info: [{"id": "d6c4d45f-8dc2-4f95-947f-9320df584daf", "address": "fa:16:3e:09:83:26", "network": {"id": "80641cc0-680b-40ff-8d1b-f5a5f736a88f", "bridge": "br-int", "label": "tempest-ServersTestJSON-666288463-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "63900312a30b4942854f552758c039cf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "26472e27-9835-4f87-ab7f-ca24dfee4e83", "external-id": "nsx-vlan-transportzone-335", "segmentation_id": 335, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd6c4d45f-8d", "ovs_interfaceid": "d6c4d45f-8dc2-4f95-947f-9320df584daf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1021.986284] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118431, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.51888} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.986742] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] ba97ce22-ea4f-430b-a4c6-08204a568436/ba97ce22-ea4f-430b-a4c6-08204a568436.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1021.987190] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: ba97ce22-ea4f-430b-a4c6-08204a568436] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1021.987637] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bca3c974-9636-4aa8-b91f-8709c12701a5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.996297] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Waiting for the task: (returnval){ [ 1021.996297] env[63531]: value = "task-1118436" [ 1021.996297] env[63531]: _type = "Task" [ 1021.996297] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.008281] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118436, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.031973] env[63531]: DEBUG oslo_concurrency.lockutils [None req-99474f63-6dfe-4906-be42-7a195dcd176b tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.009s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.033195] env[63531]: DEBUG oslo_concurrency.lockutils [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 4.353s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.115864] env[63531]: DEBUG nova.network.neutron [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1022.255394] env[63531]: DEBUG nova.network.neutron [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Updating instance_info_cache with network_info: [{"id": "b731766f-738c-4d77-a00e-98ed6c683e55", "address": "fa:16:3e:d5:26:02", "network": {"id": "0667c176-cfdf-4e43-b152-e00bdc224448", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-243715985-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "308eabafea9c4188a58a0f1c22074d2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8f580e6-1d86-41ee-9ebe-c531cb9299c6", "external-id": "nsx-vlan-transportzone-150", "segmentation_id": 150, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb731766f-73", "ovs_interfaceid": "b731766f-738c-4d77-a00e-98ed6c683e55", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1022.281934] env[63531]: DEBUG oslo_vmware.api [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]520c063c-30a6-e5f0-0ad4-25e935549c81, 'name': SearchDatastore_Task, 'duration_secs': 0.023415} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.282282] env[63531]: DEBUG oslo_concurrency.lockutils [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1022.282521] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b6db77e6-4447-4b22-93d7-265ff0ada0bd] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1022.282761] env[63531]: DEBUG oslo_concurrency.lockutils [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1022.282911] env[63531]: DEBUG oslo_concurrency.lockutils [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1022.283111] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1022.283378] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-74cc6984-5120-453d-9216-02bcaa42087a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.301136] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1022.301343] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1022.302075] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a4becb88-c35a-46f5-a3a6-d4a7915ecafd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.308805] env[63531]: DEBUG oslo_vmware.api [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 1022.308805] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5251bf12-6500-d123-326d-46841e40f046" [ 1022.308805] env[63531]: _type = "Task" [ 1022.308805] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.316791] env[63531]: DEBUG oslo_vmware.api [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5251bf12-6500-d123-326d-46841e40f046, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.466229] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118435, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.481493] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Creating Snapshot of the VM instance {{(pid=63531) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1022.481957] env[63531]: DEBUG oslo_concurrency.lockutils [req-2f3bb8d7-68dc-49a3-af62-c584e982564b req-c58376c0-e5a3-4bcd-804e-167058fb4a19 service nova] Releasing lock "refresh_cache-b6db77e6-4447-4b22-93d7-265ff0ada0bd" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1022.482338] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-3b71b471-58f9-4c5c-9674-6bed6d40e931 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.490366] env[63531]: DEBUG oslo_vmware.api [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 1022.490366] env[63531]: value = "task-1118437" [ 1022.490366] env[63531]: _type = "Task" [ 1022.490366] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.499436] env[63531]: DEBUG oslo_vmware.api [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118437, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.507501] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118436, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068306} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.507765] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: ba97ce22-ea4f-430b-a4c6-08204a568436] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1022.508596] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f11cd2bb-b4c5-40bc-99b5-f43d806b1230 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.530586] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: ba97ce22-ea4f-430b-a4c6-08204a568436] Reconfiguring VM instance instance-00000061 to attach disk [datastore2] ba97ce22-ea4f-430b-a4c6-08204a568436/ba97ce22-ea4f-430b-a4c6-08204a568436.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1022.530859] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fe8d4223-362e-4244-8b6b-2b5cd6c1a93e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.547118] env[63531]: INFO nova.compute.claims [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1022.557623] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Waiting for the task: (returnval){ [ 1022.557623] env[63531]: value = "task-1118438" [ 1022.557623] env[63531]: _type = "Task" [ 1022.557623] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.566404] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118438, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.566782] env[63531]: DEBUG oslo_concurrency.lockutils [None req-64ff1f09-13f0-4ca3-b973-cc3c91729e37 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquiring lock "911f23f7-e320-48f3-87a9-0239b013feff" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.567023] env[63531]: DEBUG oslo_concurrency.lockutils [None req-64ff1f09-13f0-4ca3-b973-cc3c91729e37 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "911f23f7-e320-48f3-87a9-0239b013feff" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.567232] env[63531]: DEBUG oslo_concurrency.lockutils [None req-64ff1f09-13f0-4ca3-b973-cc3c91729e37 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquiring lock "911f23f7-e320-48f3-87a9-0239b013feff-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.567424] env[63531]: DEBUG oslo_concurrency.lockutils [None req-64ff1f09-13f0-4ca3-b973-cc3c91729e37 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "911f23f7-e320-48f3-87a9-0239b013feff-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.567595] env[63531]: DEBUG oslo_concurrency.lockutils [None req-64ff1f09-13f0-4ca3-b973-cc3c91729e37 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "911f23f7-e320-48f3-87a9-0239b013feff-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.570020] env[63531]: INFO nova.compute.manager [None req-64ff1f09-13f0-4ca3-b973-cc3c91729e37 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Terminating instance [ 1022.571835] env[63531]: DEBUG nova.compute.manager [None req-64ff1f09-13f0-4ca3-b973-cc3c91729e37 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1022.572048] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-64ff1f09-13f0-4ca3-b973-cc3c91729e37 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1022.572808] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e229efb-f210-483c-b957-3a62a9c9a59b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.580712] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-64ff1f09-13f0-4ca3-b973-cc3c91729e37 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1022.580959] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c84c14f0-12ec-4ac0-add5-820c274ca4e2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.588086] env[63531]: DEBUG oslo_vmware.api [None req-64ff1f09-13f0-4ca3-b973-cc3c91729e37 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 1022.588086] env[63531]: value = "task-1118439" [ 1022.588086] env[63531]: _type = "Task" [ 1022.588086] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.596697] env[63531]: DEBUG oslo_vmware.api [None req-64ff1f09-13f0-4ca3-b973-cc3c91729e37 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118439, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.758573] env[63531]: DEBUG oslo_concurrency.lockutils [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Releasing lock "refresh_cache-96c2910b-1e25-4053-9ad9-c155402ef0ac" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1022.758850] env[63531]: DEBUG nova.compute.manager [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Instance network_info: |[{"id": "b731766f-738c-4d77-a00e-98ed6c683e55", "address": "fa:16:3e:d5:26:02", "network": {"id": "0667c176-cfdf-4e43-b152-e00bdc224448", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-243715985-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "308eabafea9c4188a58a0f1c22074d2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8f580e6-1d86-41ee-9ebe-c531cb9299c6", "external-id": "nsx-vlan-transportzone-150", "segmentation_id": 150, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb731766f-73", "ovs_interfaceid": "b731766f-738c-4d77-a00e-98ed6c683e55", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1022.759419] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d5:26:02', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c8f580e6-1d86-41ee-9ebe-c531cb9299c6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b731766f-738c-4d77-a00e-98ed6c683e55', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1022.767359] env[63531]: DEBUG oslo.service.loopingcall [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1022.767625] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1022.767857] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3ca18c36-7ab1-44e5-9227-53df1630df6e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.790022] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1022.790022] env[63531]: value = "task-1118440" [ 1022.790022] env[63531]: _type = "Task" [ 1022.790022] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.797908] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118440, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.818168] env[63531]: DEBUG oslo_vmware.api [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5251bf12-6500-d123-326d-46841e40f046, 'name': SearchDatastore_Task, 'duration_secs': 0.019492} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.818965] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bee77e18-3d04-4a2a-b49f-1836938ead96 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.824622] env[63531]: DEBUG oslo_vmware.api [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 1022.824622] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52a7a6bd-239e-9f86-8041-6f58f344e6e2" [ 1022.824622] env[63531]: _type = "Task" [ 1022.824622] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.835667] env[63531]: DEBUG oslo_vmware.api [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52a7a6bd-239e-9f86-8041-6f58f344e6e2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.965283] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118435, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.003228] env[63531]: DEBUG oslo_vmware.api [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118437, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.009214] env[63531]: DEBUG nova.compute.manager [req-a592c26b-bf19-4746-8c6f-22ad88d6874f req-b50c7f80-193f-4cb8-9bac-909669604f09 service nova] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Received event network-changed-b731766f-738c-4d77-a00e-98ed6c683e55 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1023.009386] env[63531]: DEBUG nova.compute.manager [req-a592c26b-bf19-4746-8c6f-22ad88d6874f req-b50c7f80-193f-4cb8-9bac-909669604f09 service nova] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Refreshing instance network info cache due to event network-changed-b731766f-738c-4d77-a00e-98ed6c683e55. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1023.009610] env[63531]: DEBUG oslo_concurrency.lockutils [req-a592c26b-bf19-4746-8c6f-22ad88d6874f req-b50c7f80-193f-4cb8-9bac-909669604f09 service nova] Acquiring lock "refresh_cache-96c2910b-1e25-4053-9ad9-c155402ef0ac" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1023.009747] env[63531]: DEBUG oslo_concurrency.lockutils [req-a592c26b-bf19-4746-8c6f-22ad88d6874f req-b50c7f80-193f-4cb8-9bac-909669604f09 service nova] Acquired lock "refresh_cache-96c2910b-1e25-4053-9ad9-c155402ef0ac" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1023.010291] env[63531]: DEBUG nova.network.neutron [req-a592c26b-bf19-4746-8c6f-22ad88d6874f req-b50c7f80-193f-4cb8-9bac-909669604f09 service nova] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Refreshing network info cache for port b731766f-738c-4d77-a00e-98ed6c683e55 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1023.054034] env[63531]: INFO nova.compute.resource_tracker [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Updating resource usage from migration 2e83b091-f95e-45ca-9edc-3ab1f0e9a94a [ 1023.071148] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118438, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.101163] env[63531]: DEBUG oslo_vmware.api [None req-64ff1f09-13f0-4ca3-b973-cc3c91729e37 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118439, 'name': PowerOffVM_Task, 'duration_secs': 0.219663} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.101444] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-64ff1f09-13f0-4ca3-b973-cc3c91729e37 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1023.101613] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-64ff1f09-13f0-4ca3-b973-cc3c91729e37 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1023.101875] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a38f5539-cc0c-4e05-bc2c-41d2ff293b37 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.217534] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-64ff1f09-13f0-4ca3-b973-cc3c91729e37 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1023.217774] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-64ff1f09-13f0-4ca3-b973-cc3c91729e37 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1023.218022] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-64ff1f09-13f0-4ca3-b973-cc3c91729e37 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Deleting the datastore file [datastore1] 911f23f7-e320-48f3-87a9-0239b013feff {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1023.218322] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9908a702-1121-4b3f-ae1f-0fffd8ff6478 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.225207] env[63531]: DEBUG oslo_vmware.api [None req-64ff1f09-13f0-4ca3-b973-cc3c91729e37 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 1023.225207] env[63531]: value = "task-1118442" [ 1023.225207] env[63531]: _type = "Task" [ 1023.225207] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.236781] env[63531]: DEBUG oslo_vmware.api [None req-64ff1f09-13f0-4ca3-b973-cc3c91729e37 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118442, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.303286] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118440, 'name': CreateVM_Task, 'duration_secs': 0.41335} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.304268] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1023.305008] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b34fa2d-b6de-49b6-932a-0e4baefb0a9c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.308114] env[63531]: DEBUG oslo_concurrency.lockutils [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1023.308343] env[63531]: DEBUG oslo_concurrency.lockutils [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1023.308673] env[63531]: DEBUG oslo_concurrency.lockutils [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1023.308933] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-10711cfd-b75e-41be-a7ef-26ec286d3b5d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.314276] env[63531]: DEBUG oslo_vmware.api [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 1023.314276] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52644223-88c5-6e32-6d70-4fb186e2ddb6" [ 1023.314276] env[63531]: _type = "Task" [ 1023.314276] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.319834] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b4e782e-dce7-4cee-a9eb-e7ae2dbacf43 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.328020] env[63531]: DEBUG oslo_vmware.api [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52644223-88c5-6e32-6d70-4fb186e2ddb6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.356849] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fe9c48f-29c8-4a98-833c-2dc1e4658057 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.364928] env[63531]: DEBUG oslo_vmware.api [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52a7a6bd-239e-9f86-8041-6f58f344e6e2, 'name': SearchDatastore_Task, 'duration_secs': 0.018787} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.367440] env[63531]: DEBUG oslo_concurrency.lockutils [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1023.367679] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] b6db77e6-4447-4b22-93d7-265ff0ada0bd/b6db77e6-4447-4b22-93d7-265ff0ada0bd.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1023.368057] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ac3664aa-141f-49f3-a10d-9e44b5d4fe69 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.371463] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17ffba33-d6f0-4f82-b3e0-9d453318dca5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.391492] env[63531]: DEBUG nova.compute.provider_tree [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1023.394087] env[63531]: DEBUG oslo_vmware.api [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 1023.394087] env[63531]: value = "task-1118443" [ 1023.394087] env[63531]: _type = "Task" [ 1023.394087] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.403937] env[63531]: DEBUG oslo_vmware.api [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118443, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.466401] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118435, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.503738] env[63531]: DEBUG oslo_vmware.api [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118437, 'name': CreateSnapshot_Task, 'duration_secs': 0.738355} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.504163] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Created Snapshot of the VM instance {{(pid=63531) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1023.505296] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-181d3211-e557-4cb7-aeef-fa0f6a52e48a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.573157] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118438, 'name': ReconfigVM_Task, 'duration_secs': 0.528629} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.573482] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: ba97ce22-ea4f-430b-a4c6-08204a568436] Reconfigured VM instance instance-00000061 to attach disk [datastore2] ba97ce22-ea4f-430b-a4c6-08204a568436/ba97ce22-ea4f-430b-a4c6-08204a568436.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1023.574172] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dce9e0bb-1887-4ca7-bb61-1b044bf88a52 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.583523] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Waiting for the task: (returnval){ [ 1023.583523] env[63531]: value = "task-1118444" [ 1023.583523] env[63531]: _type = "Task" [ 1023.583523] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.596856] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118444, 'name': Rename_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.736900] env[63531]: DEBUG oslo_vmware.api [None req-64ff1f09-13f0-4ca3-b973-cc3c91729e37 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118442, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.174149} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.737314] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-64ff1f09-13f0-4ca3-b973-cc3c91729e37 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1023.737599] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-64ff1f09-13f0-4ca3-b973-cc3c91729e37 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1023.737814] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-64ff1f09-13f0-4ca3-b973-cc3c91729e37 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1023.738100] env[63531]: INFO nova.compute.manager [None req-64ff1f09-13f0-4ca3-b973-cc3c91729e37 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Took 1.17 seconds to destroy the instance on the hypervisor. [ 1023.738400] env[63531]: DEBUG oslo.service.loopingcall [None req-64ff1f09-13f0-4ca3-b973-cc3c91729e37 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1023.738654] env[63531]: DEBUG nova.compute.manager [-] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1023.738782] env[63531]: DEBUG nova.network.neutron [-] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1023.749279] env[63531]: DEBUG nova.network.neutron [req-a592c26b-bf19-4746-8c6f-22ad88d6874f req-b50c7f80-193f-4cb8-9bac-909669604f09 service nova] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Updated VIF entry in instance network info cache for port b731766f-738c-4d77-a00e-98ed6c683e55. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1023.749664] env[63531]: DEBUG nova.network.neutron [req-a592c26b-bf19-4746-8c6f-22ad88d6874f req-b50c7f80-193f-4cb8-9bac-909669604f09 service nova] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Updating instance_info_cache with network_info: [{"id": "b731766f-738c-4d77-a00e-98ed6c683e55", "address": "fa:16:3e:d5:26:02", "network": {"id": "0667c176-cfdf-4e43-b152-e00bdc224448", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-243715985-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "308eabafea9c4188a58a0f1c22074d2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8f580e6-1d86-41ee-9ebe-c531cb9299c6", "external-id": "nsx-vlan-transportzone-150", "segmentation_id": 150, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb731766f-73", "ovs_interfaceid": "b731766f-738c-4d77-a00e-98ed6c683e55", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1023.827476] env[63531]: DEBUG oslo_vmware.api [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52644223-88c5-6e32-6d70-4fb186e2ddb6, 'name': SearchDatastore_Task, 'duration_secs': 0.014693} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.827804] env[63531]: DEBUG oslo_concurrency.lockutils [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1023.828086] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1023.828351] env[63531]: DEBUG oslo_concurrency.lockutils [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1023.828503] env[63531]: DEBUG oslo_concurrency.lockutils [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1023.828690] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1023.828985] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-49eb97f5-54c1-422a-a9e1-738c7a90a3be {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.843650] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1023.843846] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1023.844560] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c8829be8-c76f-43e5-a6a5-627b0ade1ce0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.851282] env[63531]: DEBUG oslo_vmware.api [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 1023.851282] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52927d9f-2961-2dad-13df-cb19256894d1" [ 1023.851282] env[63531]: _type = "Task" [ 1023.851282] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.861264] env[63531]: DEBUG oslo_vmware.api [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52927d9f-2961-2dad-13df-cb19256894d1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.896022] env[63531]: DEBUG nova.scheduler.client.report [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1023.907721] env[63531]: DEBUG oslo_vmware.api [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118443, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.502021} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.907988] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] b6db77e6-4447-4b22-93d7-265ff0ada0bd/b6db77e6-4447-4b22-93d7-265ff0ada0bd.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1023.908231] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b6db77e6-4447-4b22-93d7-265ff0ada0bd] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1023.908481] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c326dbeb-0bb3-42b6-89b5-ed2454d36190 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.914555] env[63531]: DEBUG oslo_vmware.api [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 1023.914555] env[63531]: value = "task-1118445" [ 1023.914555] env[63531]: _type = "Task" [ 1023.914555] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.934019] env[63531]: DEBUG oslo_vmware.api [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118445, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.969829] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118435, 'name': PowerOnVM_Task, 'duration_secs': 1.68662} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.970265] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: afa5f2c8-e519-4eb7-8e0e-ada9efb630e7] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1023.970589] env[63531]: INFO nova.compute.manager [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: afa5f2c8-e519-4eb7-8e0e-ada9efb630e7] Took 10.41 seconds to spawn the instance on the hypervisor. [ 1023.970866] env[63531]: DEBUG nova.compute.manager [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: afa5f2c8-e519-4eb7-8e0e-ada9efb630e7] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1023.972041] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b46a9153-a80c-4d65-a7b9-5d1a5b6f17e0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.028288] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Creating linked-clone VM from snapshot {{(pid=63531) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1024.029335] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-f59e860d-c1c4-48f0-87a3-0f698930c7bb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.041017] env[63531]: DEBUG oslo_vmware.api [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 1024.041017] env[63531]: value = "task-1118446" [ 1024.041017] env[63531]: _type = "Task" [ 1024.041017] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.050160] env[63531]: DEBUG oslo_vmware.api [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118446, 'name': CloneVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.094369] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118444, 'name': Rename_Task, 'duration_secs': 0.252588} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.094662] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: ba97ce22-ea4f-430b-a4c6-08204a568436] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1024.094926] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-533c2347-0892-4b8a-9a09-c43f2ffd7eb9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.103311] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Waiting for the task: (returnval){ [ 1024.103311] env[63531]: value = "task-1118447" [ 1024.103311] env[63531]: _type = "Task" [ 1024.103311] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.115625] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118447, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.210480] env[63531]: DEBUG nova.compute.manager [req-e1839c38-4882-445a-a098-6ec5e644019b req-83ab213b-da11-45fb-96ed-9bef16776245 service nova] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Received event network-vif-deleted-d3f47a79-f699-4d63-9412-724fe1f15913 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1024.210635] env[63531]: INFO nova.compute.manager [req-e1839c38-4882-445a-a098-6ec5e644019b req-83ab213b-da11-45fb-96ed-9bef16776245 service nova] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Neutron deleted interface d3f47a79-f699-4d63-9412-724fe1f15913; detaching it from the instance and deleting it from the info cache [ 1024.210815] env[63531]: DEBUG nova.network.neutron [req-e1839c38-4882-445a-a098-6ec5e644019b req-83ab213b-da11-45fb-96ed-9bef16776245 service nova] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1024.252831] env[63531]: DEBUG oslo_concurrency.lockutils [req-a592c26b-bf19-4746-8c6f-22ad88d6874f req-b50c7f80-193f-4cb8-9bac-909669604f09 service nova] Releasing lock "refresh_cache-96c2910b-1e25-4053-9ad9-c155402ef0ac" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1024.364458] env[63531]: DEBUG oslo_vmware.api [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52927d9f-2961-2dad-13df-cb19256894d1, 'name': SearchDatastore_Task, 'duration_secs': 0.020192} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.365697] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fd31ec85-3924-4528-80bc-d5de1d0de53b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.376016] env[63531]: DEBUG oslo_vmware.api [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 1024.376016] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52caa269-e1bb-50e2-ba70-451a2bcb200b" [ 1024.376016] env[63531]: _type = "Task" [ 1024.376016] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.387739] env[63531]: DEBUG oslo_vmware.api [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52caa269-e1bb-50e2-ba70-451a2bcb200b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.404627] env[63531]: DEBUG oslo_concurrency.lockutils [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.371s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.407015] env[63531]: INFO nova.compute.manager [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Migrating [ 1024.431552] env[63531]: DEBUG oslo_vmware.api [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118445, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080159} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.432036] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b6db77e6-4447-4b22-93d7-265ff0ada0bd] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1024.433383] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95a47f0e-9117-42de-bb62-781491e1b072 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.460797] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b6db77e6-4447-4b22-93d7-265ff0ada0bd] Reconfiguring VM instance instance-00000062 to attach disk [datastore1] b6db77e6-4447-4b22-93d7-265ff0ada0bd/b6db77e6-4447-4b22-93d7-265ff0ada0bd.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1024.461744] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f78a0ac8-00e8-47a7-bbff-511a2f034021 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.491023] env[63531]: DEBUG oslo_vmware.api [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 1024.491023] env[63531]: value = "task-1118448" [ 1024.491023] env[63531]: _type = "Task" [ 1024.491023] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.496300] env[63531]: INFO nova.compute.manager [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: afa5f2c8-e519-4eb7-8e0e-ada9efb630e7] Took 18.56 seconds to build instance. [ 1024.504824] env[63531]: DEBUG oslo_vmware.api [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118448, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.556902] env[63531]: DEBUG oslo_vmware.api [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118446, 'name': CloneVM_Task} progress is 94%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.613841] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118447, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.687122] env[63531]: DEBUG nova.network.neutron [-] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1024.713425] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-962f1465-bf7c-4b1d-8155-8167d29c3538 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.725049] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e2dc35d-044d-4f3c-8bf3-f5ab17dfc5f0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.765138] env[63531]: DEBUG nova.compute.manager [req-e1839c38-4882-445a-a098-6ec5e644019b req-83ab213b-da11-45fb-96ed-9bef16776245 service nova] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Detach interface failed, port_id=d3f47a79-f699-4d63-9412-724fe1f15913, reason: Instance 911f23f7-e320-48f3-87a9-0239b013feff could not be found. {{(pid=63531) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1024.884498] env[63531]: DEBUG oslo_vmware.api [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52caa269-e1bb-50e2-ba70-451a2bcb200b, 'name': SearchDatastore_Task, 'duration_secs': 0.017288} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.884794] env[63531]: DEBUG oslo_concurrency.lockutils [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1024.885071] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 96c2910b-1e25-4053-9ad9-c155402ef0ac/96c2910b-1e25-4053-9ad9-c155402ef0ac.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1024.885363] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bc12127d-16fb-4f07-aaa7-4662e0c9e791 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.891679] env[63531]: DEBUG oslo_vmware.api [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 1024.891679] env[63531]: value = "task-1118449" [ 1024.891679] env[63531]: _type = "Task" [ 1024.891679] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.901359] env[63531]: DEBUG oslo_vmware.api [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118449, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.927060] env[63531]: DEBUG oslo_concurrency.lockutils [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "refresh_cache-5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1024.927060] env[63531]: DEBUG oslo_concurrency.lockutils [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquired lock "refresh_cache-5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.927060] env[63531]: DEBUG nova.network.neutron [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1024.998627] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Lock "afa5f2c8-e519-4eb7-8e0e-ada9efb630e7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.073s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.003047] env[63531]: DEBUG oslo_vmware.api [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118448, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.052100] env[63531]: DEBUG oslo_vmware.api [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118446, 'name': CloneVM_Task} progress is 94%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.114492] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118447, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.190489] env[63531]: INFO nova.compute.manager [-] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Took 1.45 seconds to deallocate network for instance. [ 1025.405738] env[63531]: DEBUG oslo_vmware.api [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118449, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.443294} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.405738] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 96c2910b-1e25-4053-9ad9-c155402ef0ac/96c2910b-1e25-4053-9ad9-c155402ef0ac.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1025.405738] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1025.406106] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c40f1918-d4f5-445f-97e6-f0418403f91f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.412731] env[63531]: DEBUG oslo_vmware.api [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 1025.412731] env[63531]: value = "task-1118450" [ 1025.412731] env[63531]: _type = "Task" [ 1025.412731] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.421251] env[63531]: DEBUG oslo_vmware.api [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118450, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.503604] env[63531]: DEBUG oslo_vmware.api [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118448, 'name': ReconfigVM_Task, 'duration_secs': 0.707978} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.506228] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b6db77e6-4447-4b22-93d7-265ff0ada0bd] Reconfigured VM instance instance-00000062 to attach disk [datastore1] b6db77e6-4447-4b22-93d7-265ff0ada0bd/b6db77e6-4447-4b22-93d7-265ff0ada0bd.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1025.506853] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c32826fc-0728-43fb-b056-56edff6a3a41 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.514138] env[63531]: DEBUG oslo_vmware.api [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 1025.514138] env[63531]: value = "task-1118451" [ 1025.514138] env[63531]: _type = "Task" [ 1025.514138] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.521762] env[63531]: DEBUG oslo_vmware.api [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118451, 'name': Rename_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.554580] env[63531]: DEBUG oslo_vmware.api [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118446, 'name': CloneVM_Task} progress is 95%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.613852] env[63531]: DEBUG oslo_vmware.api [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118447, 'name': PowerOnVM_Task, 'duration_secs': 1.078569} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.614202] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: ba97ce22-ea4f-430b-a4c6-08204a568436] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1025.614454] env[63531]: INFO nova.compute.manager [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: ba97ce22-ea4f-430b-a4c6-08204a568436] Took 9.69 seconds to spawn the instance on the hypervisor. [ 1025.614659] env[63531]: DEBUG nova.compute.manager [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: ba97ce22-ea4f-430b-a4c6-08204a568436] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1025.615429] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c53fac1f-0216-4808-9d67-4d9c74123f20 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.702685] env[63531]: DEBUG oslo_concurrency.lockutils [None req-64ff1f09-13f0-4ca3-b973-cc3c91729e37 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.703036] env[63531]: DEBUG oslo_concurrency.lockutils [None req-64ff1f09-13f0-4ca3-b973-cc3c91729e37 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.706935] env[63531]: DEBUG nova.objects.instance [None req-64ff1f09-13f0-4ca3-b973-cc3c91729e37 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lazy-loading 'resources' on Instance uuid 911f23f7-e320-48f3-87a9-0239b013feff {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1025.784063] env[63531]: DEBUG nova.network.neutron [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Updating instance_info_cache with network_info: [{"id": "91a7c60f-cbcb-4fe3-9bd0-69337882b7ce", "address": "fa:16:3e:46:61:c6", "network": {"id": "bf99b8cb-9efb-4a74-9db1-0f5f0875f26a", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-998205595-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2e99349303e413b9f7896d449cb7dff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc6e6fe1-c4f5-4389-a49f-0978060eebb4", "external-id": "nsx-vlan-transportzone-829", "segmentation_id": 829, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91a7c60f-cb", "ovs_interfaceid": "91a7c60f-cbcb-4fe3-9bd0-69337882b7ce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1025.927963] env[63531]: DEBUG oslo_vmware.api [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118450, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065343} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.928356] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1025.929265] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f321347-75e6-45b3-8085-d230a5bced6e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.951788] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] 96c2910b-1e25-4053-9ad9-c155402ef0ac/96c2910b-1e25-4053-9ad9-c155402ef0ac.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1025.952095] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6d86dcac-9166-4811-b978-5c9782cadb73 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.971030] env[63531]: DEBUG oslo_vmware.api [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 1025.971030] env[63531]: value = "task-1118452" [ 1025.971030] env[63531]: _type = "Task" [ 1025.971030] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.978303] env[63531]: DEBUG oslo_vmware.api [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118452, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.023745] env[63531]: DEBUG oslo_vmware.api [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118451, 'name': Rename_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.052204] env[63531]: DEBUG oslo_vmware.api [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118446, 'name': CloneVM_Task, 'duration_secs': 1.701712} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.052485] env[63531]: INFO nova.virt.vmwareapi.vmops [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Created linked-clone VM from snapshot [ 1026.053206] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b87a4a8a-6f58-42a7-8ef9-8380eef7246d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.062162] env[63531]: DEBUG nova.virt.vmwareapi.images [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Uploading image 87aadec8-cd66-47c5-a3d7-3ae819edebf8 {{(pid=63531) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1026.087739] env[63531]: DEBUG oslo_vmware.rw_handles [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1026.087739] env[63531]: value = "vm-244787" [ 1026.087739] env[63531]: _type = "VirtualMachine" [ 1026.087739] env[63531]: }. {{(pid=63531) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1026.088137] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-75c0ed2a-42e9-427a-bdc0-d2e56147b705 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.094766] env[63531]: DEBUG oslo_vmware.rw_handles [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lease: (returnval){ [ 1026.094766] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]523d1c76-ff59-cf59-3363-441cfeacfe24" [ 1026.094766] env[63531]: _type = "HttpNfcLease" [ 1026.094766] env[63531]: } obtained for exporting VM: (result){ [ 1026.094766] env[63531]: value = "vm-244787" [ 1026.094766] env[63531]: _type = "VirtualMachine" [ 1026.094766] env[63531]: }. {{(pid=63531) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1026.095056] env[63531]: DEBUG oslo_vmware.api [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the lease: (returnval){ [ 1026.095056] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]523d1c76-ff59-cf59-3363-441cfeacfe24" [ 1026.095056] env[63531]: _type = "HttpNfcLease" [ 1026.095056] env[63531]: } to be ready. {{(pid=63531) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1026.101492] env[63531]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1026.101492] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]523d1c76-ff59-cf59-3363-441cfeacfe24" [ 1026.101492] env[63531]: _type = "HttpNfcLease" [ 1026.101492] env[63531]: } is initializing. {{(pid=63531) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1026.130287] env[63531]: INFO nova.compute.manager [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: ba97ce22-ea4f-430b-a4c6-08204a568436] Took 20.17 seconds to build instance. [ 1026.286522] env[63531]: DEBUG oslo_concurrency.lockutils [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Releasing lock "refresh_cache-5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1026.417923] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8312109-65e5-4d0a-9e08-43608c228ba9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.425377] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a427136a-3fbd-498c-929f-562ad2527c4a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.467618] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c048872-d52c-423b-97a9-14101cd1f746 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.477970] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5211ec33-fb8d-4947-9786-9bf00732b9fa {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.486353] env[63531]: DEBUG oslo_vmware.api [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118452, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.495666] env[63531]: DEBUG nova.compute.provider_tree [None req-64ff1f09-13f0-4ca3-b973-cc3c91729e37 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1026.523178] env[63531]: DEBUG oslo_vmware.api [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118451, 'name': Rename_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.601930] env[63531]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1026.601930] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]523d1c76-ff59-cf59-3363-441cfeacfe24" [ 1026.601930] env[63531]: _type = "HttpNfcLease" [ 1026.601930] env[63531]: } is ready. {{(pid=63531) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1026.602323] env[63531]: DEBUG oslo_vmware.rw_handles [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1026.602323] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]523d1c76-ff59-cf59-3363-441cfeacfe24" [ 1026.602323] env[63531]: _type = "HttpNfcLease" [ 1026.602323] env[63531]: }. {{(pid=63531) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1026.602931] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e98c47da-0f6d-4f28-8abc-8f2bb91ede6c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.609885] env[63531]: DEBUG oslo_vmware.rw_handles [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e47cd7-a2be-aef4-7628-cbce47f138d1/disk-0.vmdk from lease info. {{(pid=63531) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1026.610148] env[63531]: DEBUG oslo_vmware.rw_handles [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e47cd7-a2be-aef4-7628-cbce47f138d1/disk-0.vmdk for reading. {{(pid=63531) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1026.665950] env[63531]: DEBUG oslo_concurrency.lockutils [None req-89655885-53ad-493a-9114-f2b1c23800d2 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Acquiring lock "afa5f2c8-e519-4eb7-8e0e-ada9efb630e7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.666283] env[63531]: DEBUG oslo_concurrency.lockutils [None req-89655885-53ad-493a-9114-f2b1c23800d2 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Lock "afa5f2c8-e519-4eb7-8e0e-ada9efb630e7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.666501] env[63531]: DEBUG oslo_concurrency.lockutils [None req-89655885-53ad-493a-9114-f2b1c23800d2 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Acquiring lock "afa5f2c8-e519-4eb7-8e0e-ada9efb630e7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.666688] env[63531]: DEBUG oslo_concurrency.lockutils [None req-89655885-53ad-493a-9114-f2b1c23800d2 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Lock "afa5f2c8-e519-4eb7-8e0e-ada9efb630e7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.666860] env[63531]: DEBUG oslo_concurrency.lockutils [None req-89655885-53ad-493a-9114-f2b1c23800d2 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Lock "afa5f2c8-e519-4eb7-8e0e-ada9efb630e7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.668649] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5f6aee5f-a17b-48fb-a403-a3f77f6134c0 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Lock "ba97ce22-ea4f-430b-a4c6-08204a568436" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.710s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.670391] env[63531]: INFO nova.compute.manager [None req-89655885-53ad-493a-9114-f2b1c23800d2 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: afa5f2c8-e519-4eb7-8e0e-ada9efb630e7] Terminating instance [ 1026.672269] env[63531]: DEBUG nova.compute.manager [None req-89655885-53ad-493a-9114-f2b1c23800d2 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: afa5f2c8-e519-4eb7-8e0e-ada9efb630e7] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1026.672470] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-89655885-53ad-493a-9114-f2b1c23800d2 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: afa5f2c8-e519-4eb7-8e0e-ada9efb630e7] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1026.673313] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11b2d544-6960-44e3-8f43-0e5060b72605 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.680321] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-89655885-53ad-493a-9114-f2b1c23800d2 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: afa5f2c8-e519-4eb7-8e0e-ada9efb630e7] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1026.680565] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3a3c9787-b2b4-419b-b950-e35ad71f7304 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.687024] env[63531]: DEBUG oslo_vmware.api [None req-89655885-53ad-493a-9114-f2b1c23800d2 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Waiting for the task: (returnval){ [ 1026.687024] env[63531]: value = "task-1118454" [ 1026.687024] env[63531]: _type = "Task" [ 1026.687024] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.690491] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f1609b7f-e5ee-4b90-9864-431682b40af5 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Acquiring lock "ba97ce22-ea4f-430b-a4c6-08204a568436" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.690705] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f1609b7f-e5ee-4b90-9864-431682b40af5 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Lock "ba97ce22-ea4f-430b-a4c6-08204a568436" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.690900] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f1609b7f-e5ee-4b90-9864-431682b40af5 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Acquiring lock "ba97ce22-ea4f-430b-a4c6-08204a568436-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.691093] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f1609b7f-e5ee-4b90-9864-431682b40af5 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Lock "ba97ce22-ea4f-430b-a4c6-08204a568436-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.691265] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f1609b7f-e5ee-4b90-9864-431682b40af5 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Lock "ba97ce22-ea4f-430b-a4c6-08204a568436-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.693259] env[63531]: INFO nova.compute.manager [None req-f1609b7f-e5ee-4b90-9864-431682b40af5 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: ba97ce22-ea4f-430b-a4c6-08204a568436] Terminating instance [ 1026.697418] env[63531]: DEBUG oslo_vmware.api [None req-89655885-53ad-493a-9114-f2b1c23800d2 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118454, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.697974] env[63531]: DEBUG nova.compute.manager [None req-f1609b7f-e5ee-4b90-9864-431682b40af5 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: ba97ce22-ea4f-430b-a4c6-08204a568436] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1026.698200] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f1609b7f-e5ee-4b90-9864-431682b40af5 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: ba97ce22-ea4f-430b-a4c6-08204a568436] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1026.698928] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f0ddf0b-e104-4320-a19b-4060da1649cf {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.705271] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1609b7f-e5ee-4b90-9864-431682b40af5 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: ba97ce22-ea4f-430b-a4c6-08204a568436] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1026.705488] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-275cc3cd-f008-4498-b80e-ed4fff2aa142 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.710681] env[63531]: DEBUG oslo_vmware.api [None req-f1609b7f-e5ee-4b90-9864-431682b40af5 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Waiting for the task: (returnval){ [ 1026.710681] env[63531]: value = "task-1118455" [ 1026.710681] env[63531]: _type = "Task" [ 1026.710681] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.718107] env[63531]: DEBUG oslo_vmware.api [None req-f1609b7f-e5ee-4b90-9864-431682b40af5 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118455, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.750787] env[63531]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-7c648ebe-c88a-4696-8099-77485792156f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.981646] env[63531]: DEBUG oslo_vmware.api [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118452, 'name': ReconfigVM_Task, 'duration_secs': 0.816536} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.981952] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Reconfigured VM instance instance-00000063 to attach disk [datastore1] 96c2910b-1e25-4053-9ad9-c155402ef0ac/96c2910b-1e25-4053-9ad9-c155402ef0ac.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1026.982650] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-df831cbc-bedf-47c9-9938-25e6335d8a60 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.989942] env[63531]: DEBUG oslo_vmware.api [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 1026.989942] env[63531]: value = "task-1118456" [ 1026.989942] env[63531]: _type = "Task" [ 1026.989942] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.999549] env[63531]: DEBUG nova.scheduler.client.report [None req-64ff1f09-13f0-4ca3-b973-cc3c91729e37 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1027.002733] env[63531]: DEBUG oslo_vmware.api [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118456, 'name': Rename_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.025031] env[63531]: DEBUG oslo_vmware.api [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118451, 'name': Rename_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.197207] env[63531]: DEBUG oslo_vmware.api [None req-89655885-53ad-493a-9114-f2b1c23800d2 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118454, 'name': PowerOffVM_Task, 'duration_secs': 0.503362} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.197651] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-89655885-53ad-493a-9114-f2b1c23800d2 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: afa5f2c8-e519-4eb7-8e0e-ada9efb630e7] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1027.197897] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-89655885-53ad-493a-9114-f2b1c23800d2 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: afa5f2c8-e519-4eb7-8e0e-ada9efb630e7] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1027.198300] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5357f0e4-adff-459e-bd85-31b3432aa0fc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.221988] env[63531]: DEBUG oslo_vmware.api [None req-f1609b7f-e5ee-4b90-9864-431682b40af5 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118455, 'name': PowerOffVM_Task, 'duration_secs': 0.188738} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.222359] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1609b7f-e5ee-4b90-9864-431682b40af5 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: ba97ce22-ea4f-430b-a4c6-08204a568436] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1027.222561] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f1609b7f-e5ee-4b90-9864-431682b40af5 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: ba97ce22-ea4f-430b-a4c6-08204a568436] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1027.222935] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-62d0fe2b-737a-45f3-b1b0-4861dded01b3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.268810] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-89655885-53ad-493a-9114-f2b1c23800d2 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: afa5f2c8-e519-4eb7-8e0e-ada9efb630e7] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1027.269367] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-89655885-53ad-493a-9114-f2b1c23800d2 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: afa5f2c8-e519-4eb7-8e0e-ada9efb630e7] Deleting contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1027.269703] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-89655885-53ad-493a-9114-f2b1c23800d2 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Deleting the datastore file [datastore2] afa5f2c8-e519-4eb7-8e0e-ada9efb630e7 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1027.270090] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-20fa7c94-9710-4fe5-9511-90705fe39d30 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.277115] env[63531]: DEBUG oslo_vmware.api [None req-89655885-53ad-493a-9114-f2b1c23800d2 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Waiting for the task: (returnval){ [ 1027.277115] env[63531]: value = "task-1118459" [ 1027.277115] env[63531]: _type = "Task" [ 1027.277115] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.286126] env[63531]: DEBUG oslo_vmware.api [None req-89655885-53ad-493a-9114-f2b1c23800d2 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118459, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.291507] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f1609b7f-e5ee-4b90-9864-431682b40af5 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: ba97ce22-ea4f-430b-a4c6-08204a568436] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1027.291839] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f1609b7f-e5ee-4b90-9864-431682b40af5 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: ba97ce22-ea4f-430b-a4c6-08204a568436] Deleting contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1027.292100] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1609b7f-e5ee-4b90-9864-431682b40af5 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Deleting the datastore file [datastore2] ba97ce22-ea4f-430b-a4c6-08204a568436 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1027.292437] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ceaf9016-2bf7-49f1-a512-b7b4122a5654 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.301111] env[63531]: DEBUG oslo_vmware.api [None req-f1609b7f-e5ee-4b90-9864-431682b40af5 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Waiting for the task: (returnval){ [ 1027.301111] env[63531]: value = "task-1118460" [ 1027.301111] env[63531]: _type = "Task" [ 1027.301111] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.310451] env[63531]: DEBUG oslo_vmware.api [None req-f1609b7f-e5ee-4b90-9864-431682b40af5 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118460, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.500589] env[63531]: DEBUG oslo_vmware.api [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118456, 'name': Rename_Task, 'duration_secs': 0.149188} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.501081] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1027.501402] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e3bc2d2b-d00e-407e-b752-368ee4be49b4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.503841] env[63531]: DEBUG oslo_concurrency.lockutils [None req-64ff1f09-13f0-4ca3-b973-cc3c91729e37 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.801s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.508641] env[63531]: DEBUG oslo_vmware.api [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 1027.508641] env[63531]: value = "task-1118461" [ 1027.508641] env[63531]: _type = "Task" [ 1027.508641] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.516653] env[63531]: DEBUG oslo_vmware.api [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118461, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.524087] env[63531]: DEBUG oslo_vmware.api [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118451, 'name': Rename_Task, 'duration_secs': 1.53244} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.525093] env[63531]: INFO nova.scheduler.client.report [None req-64ff1f09-13f0-4ca3-b973-cc3c91729e37 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Deleted allocations for instance 911f23f7-e320-48f3-87a9-0239b013feff [ 1027.528712] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b6db77e6-4447-4b22-93d7-265ff0ada0bd] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1027.529312] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f54bcb16-f2ea-44c2-a4c0-624832f7174f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.536483] env[63531]: DEBUG oslo_vmware.api [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 1027.536483] env[63531]: value = "task-1118462" [ 1027.536483] env[63531]: _type = "Task" [ 1027.536483] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.547150] env[63531]: DEBUG oslo_vmware.api [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118462, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.787869] env[63531]: DEBUG oslo_vmware.api [None req-89655885-53ad-493a-9114-f2b1c23800d2 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118459, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.184557} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.788404] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-89655885-53ad-493a-9114-f2b1c23800d2 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1027.788721] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-89655885-53ad-493a-9114-f2b1c23800d2 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: afa5f2c8-e519-4eb7-8e0e-ada9efb630e7] Deleted contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1027.789133] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-89655885-53ad-493a-9114-f2b1c23800d2 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: afa5f2c8-e519-4eb7-8e0e-ada9efb630e7] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1027.789169] env[63531]: INFO nova.compute.manager [None req-89655885-53ad-493a-9114-f2b1c23800d2 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: afa5f2c8-e519-4eb7-8e0e-ada9efb630e7] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1027.789514] env[63531]: DEBUG oslo.service.loopingcall [None req-89655885-53ad-493a-9114-f2b1c23800d2 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1027.789799] env[63531]: DEBUG nova.compute.manager [-] [instance: afa5f2c8-e519-4eb7-8e0e-ada9efb630e7] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1027.789927] env[63531]: DEBUG nova.network.neutron [-] [instance: afa5f2c8-e519-4eb7-8e0e-ada9efb630e7] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1027.802032] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-628af8cc-6261-41b7-95bf-6639ceeaad89 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.824736] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Updating instance '5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7' progress to 0 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1027.834849] env[63531]: DEBUG oslo_vmware.api [None req-f1609b7f-e5ee-4b90-9864-431682b40af5 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Task: {'id': task-1118460, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.208397} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.835802] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1609b7f-e5ee-4b90-9864-431682b40af5 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1027.836014] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f1609b7f-e5ee-4b90-9864-431682b40af5 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: ba97ce22-ea4f-430b-a4c6-08204a568436] Deleted contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1027.836487] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f1609b7f-e5ee-4b90-9864-431682b40af5 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: ba97ce22-ea4f-430b-a4c6-08204a568436] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1027.836918] env[63531]: INFO nova.compute.manager [None req-f1609b7f-e5ee-4b90-9864-431682b40af5 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] [instance: ba97ce22-ea4f-430b-a4c6-08204a568436] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1027.837284] env[63531]: DEBUG oslo.service.loopingcall [None req-f1609b7f-e5ee-4b90-9864-431682b40af5 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1027.837553] env[63531]: DEBUG nova.compute.manager [-] [instance: ba97ce22-ea4f-430b-a4c6-08204a568436] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1027.837704] env[63531]: DEBUG nova.network.neutron [-] [instance: ba97ce22-ea4f-430b-a4c6-08204a568436] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1028.019521] env[63531]: DEBUG oslo_vmware.api [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118461, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.035291] env[63531]: DEBUG oslo_concurrency.lockutils [None req-64ff1f09-13f0-4ca3-b973-cc3c91729e37 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "911f23f7-e320-48f3-87a9-0239b013feff" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.468s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.047255] env[63531]: DEBUG oslo_vmware.api [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118462, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.187306] env[63531]: DEBUG nova.compute.manager [req-9a562c5e-95d3-401a-978f-2e257ccb8c42 req-62e0a800-cccb-4f99-9849-fc8aab9bdea2 service nova] [instance: ba97ce22-ea4f-430b-a4c6-08204a568436] Received event network-vif-deleted-8baef0c0-0988-4a88-8808-34fde881d31d {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1028.188139] env[63531]: INFO nova.compute.manager [req-9a562c5e-95d3-401a-978f-2e257ccb8c42 req-62e0a800-cccb-4f99-9849-fc8aab9bdea2 service nova] [instance: ba97ce22-ea4f-430b-a4c6-08204a568436] Neutron deleted interface 8baef0c0-0988-4a88-8808-34fde881d31d; detaching it from the instance and deleting it from the info cache [ 1028.188428] env[63531]: DEBUG nova.network.neutron [req-9a562c5e-95d3-401a-978f-2e257ccb8c42 req-62e0a800-cccb-4f99-9849-fc8aab9bdea2 service nova] [instance: ba97ce22-ea4f-430b-a4c6-08204a568436] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1028.206896] env[63531]: DEBUG nova.compute.manager [req-ee82a3d3-200e-43d3-b551-2f0445cfe727 req-48cd034d-6381-48dc-8437-ef50a0db816c service nova] [instance: afa5f2c8-e519-4eb7-8e0e-ada9efb630e7] Received event network-vif-deleted-b320eed5-1bc8-4774-9abd-e8cfcd640e56 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1028.207229] env[63531]: INFO nova.compute.manager [req-ee82a3d3-200e-43d3-b551-2f0445cfe727 req-48cd034d-6381-48dc-8437-ef50a0db816c service nova] [instance: afa5f2c8-e519-4eb7-8e0e-ada9efb630e7] Neutron deleted interface b320eed5-1bc8-4774-9abd-e8cfcd640e56; detaching it from the instance and deleting it from the info cache [ 1028.207494] env[63531]: DEBUG nova.network.neutron [req-ee82a3d3-200e-43d3-b551-2f0445cfe727 req-48cd034d-6381-48dc-8437-ef50a0db816c service nova] [instance: afa5f2c8-e519-4eb7-8e0e-ada9efb630e7] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1028.338519] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1028.338755] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a7ed1568-1181-4a2e-b1c7-5639a0ff714c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.347035] env[63531]: DEBUG oslo_vmware.api [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 1028.347035] env[63531]: value = "task-1118463" [ 1028.347035] env[63531]: _type = "Task" [ 1028.347035] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.356418] env[63531]: DEBUG oslo_vmware.api [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118463, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.522897] env[63531]: DEBUG oslo_vmware.api [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118461, 'name': PowerOnVM_Task, 'duration_secs': 0.803736} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.523539] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1028.523887] env[63531]: INFO nova.compute.manager [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Took 7.87 seconds to spawn the instance on the hypervisor. [ 1028.524517] env[63531]: DEBUG nova.compute.manager [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1028.525748] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86cce70d-87cc-4e15-a60e-1f938cc67a20 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.548659] env[63531]: DEBUG oslo_vmware.api [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118462, 'name': PowerOnVM_Task, 'duration_secs': 0.645878} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.548942] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b6db77e6-4447-4b22-93d7-265ff0ada0bd] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1028.549190] env[63531]: INFO nova.compute.manager [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b6db77e6-4447-4b22-93d7-265ff0ada0bd] Took 10.31 seconds to spawn the instance on the hypervisor. [ 1028.549416] env[63531]: DEBUG nova.compute.manager [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b6db77e6-4447-4b22-93d7-265ff0ada0bd] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1028.550328] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5ed4f50-0819-4ca5-ae1c-75f7aed2cdc4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.592680] env[63531]: DEBUG nova.network.neutron [-] [instance: ba97ce22-ea4f-430b-a4c6-08204a568436] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1028.626813] env[63531]: DEBUG nova.network.neutron [-] [instance: afa5f2c8-e519-4eb7-8e0e-ada9efb630e7] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1028.692015] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3c1042c5-0234-4274-8dca-f4f81cbc4f5e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.703183] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2a767ba-c294-4fe8-bd48-cdd82fc15fe8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.714827] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-84389fec-b120-488b-b5ef-dc871e6aab57 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.725582] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3ee6935-2dfb-47b1-b0ea-b2a1ae0ce8fa {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.748170] env[63531]: DEBUG nova.compute.manager [req-9a562c5e-95d3-401a-978f-2e257ccb8c42 req-62e0a800-cccb-4f99-9849-fc8aab9bdea2 service nova] [instance: ba97ce22-ea4f-430b-a4c6-08204a568436] Detach interface failed, port_id=8baef0c0-0988-4a88-8808-34fde881d31d, reason: Instance ba97ce22-ea4f-430b-a4c6-08204a568436 could not be found. {{(pid=63531) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1028.771434] env[63531]: DEBUG nova.compute.manager [req-ee82a3d3-200e-43d3-b551-2f0445cfe727 req-48cd034d-6381-48dc-8437-ef50a0db816c service nova] [instance: afa5f2c8-e519-4eb7-8e0e-ada9efb630e7] Detach interface failed, port_id=b320eed5-1bc8-4774-9abd-e8cfcd640e56, reason: Instance afa5f2c8-e519-4eb7-8e0e-ada9efb630e7 could not be found. {{(pid=63531) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1028.857416] env[63531]: DEBUG oslo_vmware.api [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118463, 'name': PowerOffVM_Task, 'duration_secs': 0.290394} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.857732] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1028.857958] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Updating instance '5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7' progress to 17 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1028.941842] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c12caf0f-2eba-4205-bb3b-92d592445017 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquiring lock "ccf00b09-29fe-4a76-a8af-97774acd77de" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.942146] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c12caf0f-2eba-4205-bb3b-92d592445017 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Lock "ccf00b09-29fe-4a76-a8af-97774acd77de" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.048823] env[63531]: INFO nova.compute.manager [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Took 16.68 seconds to build instance. [ 1029.067536] env[63531]: INFO nova.compute.manager [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b6db77e6-4447-4b22-93d7-265ff0ada0bd] Took 19.58 seconds to build instance. [ 1029.095366] env[63531]: INFO nova.compute.manager [-] [instance: ba97ce22-ea4f-430b-a4c6-08204a568436] Took 1.26 seconds to deallocate network for instance. [ 1029.129348] env[63531]: INFO nova.compute.manager [-] [instance: afa5f2c8-e519-4eb7-8e0e-ada9efb630e7] Took 1.34 seconds to deallocate network for instance. [ 1029.365521] env[63531]: DEBUG nova.virt.hardware [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1029.365777] env[63531]: DEBUG nova.virt.hardware [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1029.365936] env[63531]: DEBUG nova.virt.hardware [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1029.366250] env[63531]: DEBUG nova.virt.hardware [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1029.366349] env[63531]: DEBUG nova.virt.hardware [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1029.370022] env[63531]: DEBUG nova.virt.hardware [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1029.370022] env[63531]: DEBUG nova.virt.hardware [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1029.370022] env[63531]: DEBUG nova.virt.hardware [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1029.370022] env[63531]: DEBUG nova.virt.hardware [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1029.370022] env[63531]: DEBUG nova.virt.hardware [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1029.370022] env[63531]: DEBUG nova.virt.hardware [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1029.373509] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-acdc9efe-9623-41cf-bb80-3d2518fbb8e9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.389776] env[63531]: DEBUG oslo_vmware.api [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 1029.389776] env[63531]: value = "task-1118464" [ 1029.389776] env[63531]: _type = "Task" [ 1029.389776] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.398892] env[63531]: DEBUG oslo_vmware.api [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118464, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.446767] env[63531]: INFO nova.compute.manager [None req-c12caf0f-2eba-4205-bb3b-92d592445017 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Detaching volume 0bd45432-88d2-467a-aacf-1a9930b4e929 [ 1029.493133] env[63531]: INFO nova.virt.block_device [None req-c12caf0f-2eba-4205-bb3b-92d592445017 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Attempting to driver detach volume 0bd45432-88d2-467a-aacf-1a9930b4e929 from mountpoint /dev/sdb [ 1029.493133] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-c12caf0f-2eba-4205-bb3b-92d592445017 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Volume detach. Driver type: vmdk {{(pid=63531) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1029.493133] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-c12caf0f-2eba-4205-bb3b-92d592445017 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244767', 'volume_id': '0bd45432-88d2-467a-aacf-1a9930b4e929', 'name': 'volume-0bd45432-88d2-467a-aacf-1a9930b4e929', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ccf00b09-29fe-4a76-a8af-97774acd77de', 'attached_at': '', 'detached_at': '', 'volume_id': '0bd45432-88d2-467a-aacf-1a9930b4e929', 'serial': '0bd45432-88d2-467a-aacf-1a9930b4e929'} {{(pid=63531) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1029.493133] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6de73fdc-34a6-4c87-846e-379588bf9f94 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.522161] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc699807-e4f9-4ca4-a5a0-5bf0199e6599 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.527961] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3df57392-9f83-4197-824d-2fe8e051d431 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.551551] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc80e646-0ea4-44cb-9d35-6d3e877098c3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.554596] env[63531]: DEBUG oslo_concurrency.lockutils [None req-206bc135-c6b0-48a2-bb01-b3e0c8cee156 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "96c2910b-1e25-4053-9ad9-c155402ef0ac" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.192s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.568512] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-c12caf0f-2eba-4205-bb3b-92d592445017 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] The volume has not been displaced from its original location: [datastore2] volume-0bd45432-88d2-467a-aacf-1a9930b4e929/volume-0bd45432-88d2-467a-aacf-1a9930b4e929.vmdk. No consolidation needed. {{(pid=63531) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1029.573930] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-c12caf0f-2eba-4205-bb3b-92d592445017 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Reconfiguring VM instance instance-00000051 to detach disk 2001 {{(pid=63531) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1029.574493] env[63531]: DEBUG oslo_concurrency.lockutils [None req-fc9f6c1f-c615-49e0-b2fb-e453310cda0e tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "b6db77e6-4447-4b22-93d7-265ff0ada0bd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.099s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.574777] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d86949d8-78d0-4201-9ca7-f08a2d1a60b7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.593125] env[63531]: DEBUG oslo_vmware.api [None req-c12caf0f-2eba-4205-bb3b-92d592445017 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 1029.593125] env[63531]: value = "task-1118465" [ 1029.593125] env[63531]: _type = "Task" [ 1029.593125] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.603127] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f1609b7f-e5ee-4b90-9864-431682b40af5 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.603408] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f1609b7f-e5ee-4b90-9864-431682b40af5 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.603634] env[63531]: DEBUG nova.objects.instance [None req-f1609b7f-e5ee-4b90-9864-431682b40af5 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Lazy-loading 'resources' on Instance uuid ba97ce22-ea4f-430b-a4c6-08204a568436 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1029.605153] env[63531]: DEBUG oslo_vmware.api [None req-c12caf0f-2eba-4205-bb3b-92d592445017 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118465, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.636329] env[63531]: DEBUG oslo_concurrency.lockutils [None req-89655885-53ad-493a-9114-f2b1c23800d2 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.906347] env[63531]: DEBUG oslo_vmware.api [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118464, 'name': ReconfigVM_Task, 'duration_secs': 0.27636} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.906946] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Updating instance '5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7' progress to 33 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1030.099222] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquiring lock "c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1030.099222] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1030.113448] env[63531]: DEBUG oslo_vmware.api [None req-c12caf0f-2eba-4205-bb3b-92d592445017 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118465, 'name': ReconfigVM_Task, 'duration_secs': 0.253346} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.114374] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-c12caf0f-2eba-4205-bb3b-92d592445017 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Reconfigured VM instance instance-00000051 to detach disk 2001 {{(pid=63531) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1030.120081] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-49523514-d01b-479e-ba81-0ea7ea23ce53 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.141494] env[63531]: DEBUG oslo_vmware.api [None req-c12caf0f-2eba-4205-bb3b-92d592445017 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 1030.141494] env[63531]: value = "task-1118466" [ 1030.141494] env[63531]: _type = "Task" [ 1030.141494] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.156869] env[63531]: DEBUG oslo_vmware.api [None req-c12caf0f-2eba-4205-bb3b-92d592445017 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118466, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.383783] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ea3caab-bd02-4ddc-b1cf-a01b1cbf0311 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.392029] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3fdc4e1-c9ae-4f4d-9618-e681908339b6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.428204] env[63531]: DEBUG nova.virt.hardware [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1030.428480] env[63531]: DEBUG nova.virt.hardware [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1030.428663] env[63531]: DEBUG nova.virt.hardware [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1030.428820] env[63531]: DEBUG nova.virt.hardware [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1030.428981] env[63531]: DEBUG nova.virt.hardware [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1030.429190] env[63531]: DEBUG nova.virt.hardware [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1030.429429] env[63531]: DEBUG nova.virt.hardware [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1030.429601] env[63531]: DEBUG nova.virt.hardware [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1030.429773] env[63531]: DEBUG nova.virt.hardware [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1030.429942] env[63531]: DEBUG nova.virt.hardware [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1030.430137] env[63531]: DEBUG nova.virt.hardware [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1030.435691] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Reconfiguring VM instance instance-0000005e to detach disk 2000 {{(pid=63531) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1030.437398] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a1a99c02-d6a3-4c7a-96bf-ece7cad9ce84 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.451362] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97eae606-1e8a-4957-81ec-2ae6add5dba1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.455373] env[63531]: DEBUG nova.compute.manager [req-2cae8d63-c355-4bd4-ac4a-0c5bb158463a req-57d2e6c2-2776-499b-b6bf-fd5e430702d7 service nova] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Received event network-changed-b731766f-738c-4d77-a00e-98ed6c683e55 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1030.455563] env[63531]: DEBUG nova.compute.manager [req-2cae8d63-c355-4bd4-ac4a-0c5bb158463a req-57d2e6c2-2776-499b-b6bf-fd5e430702d7 service nova] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Refreshing instance network info cache due to event network-changed-b731766f-738c-4d77-a00e-98ed6c683e55. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1030.455779] env[63531]: DEBUG oslo_concurrency.lockutils [req-2cae8d63-c355-4bd4-ac4a-0c5bb158463a req-57d2e6c2-2776-499b-b6bf-fd5e430702d7 service nova] Acquiring lock "refresh_cache-96c2910b-1e25-4053-9ad9-c155402ef0ac" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1030.455967] env[63531]: DEBUG oslo_concurrency.lockutils [req-2cae8d63-c355-4bd4-ac4a-0c5bb158463a req-57d2e6c2-2776-499b-b6bf-fd5e430702d7 service nova] Acquired lock "refresh_cache-96c2910b-1e25-4053-9ad9-c155402ef0ac" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.456096] env[63531]: DEBUG nova.network.neutron [req-2cae8d63-c355-4bd4-ac4a-0c5bb158463a req-57d2e6c2-2776-499b-b6bf-fd5e430702d7 service nova] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Refreshing network info cache for port b731766f-738c-4d77-a00e-98ed6c683e55 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1030.465966] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3803ba7-874b-4f27-a8e2-9cc3be01ffae {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.471116] env[63531]: DEBUG oslo_vmware.api [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 1030.471116] env[63531]: value = "task-1118467" [ 1030.471116] env[63531]: _type = "Task" [ 1030.471116] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.485110] env[63531]: DEBUG nova.compute.provider_tree [None req-f1609b7f-e5ee-4b90-9864-431682b40af5 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1030.492028] env[63531]: DEBUG oslo_vmware.api [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118467, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.605654] env[63531]: DEBUG nova.compute.manager [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1030.659310] env[63531]: DEBUG oslo_vmware.api [None req-c12caf0f-2eba-4205-bb3b-92d592445017 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118466, 'name': ReconfigVM_Task, 'duration_secs': 0.166728} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.660598] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-c12caf0f-2eba-4205-bb3b-92d592445017 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244767', 'volume_id': '0bd45432-88d2-467a-aacf-1a9930b4e929', 'name': 'volume-0bd45432-88d2-467a-aacf-1a9930b4e929', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ccf00b09-29fe-4a76-a8af-97774acd77de', 'attached_at': '', 'detached_at': '', 'volume_id': '0bd45432-88d2-467a-aacf-1a9930b4e929', 'serial': '0bd45432-88d2-467a-aacf-1a9930b4e929'} {{(pid=63531) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1030.808892] env[63531]: DEBUG oslo_concurrency.lockutils [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "97d3f108-e299-4ef1-84ca-532cd64905d5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1030.809233] env[63531]: DEBUG oslo_concurrency.lockutils [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "97d3f108-e299-4ef1-84ca-532cd64905d5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1030.984633] env[63531]: DEBUG oslo_vmware.api [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118467, 'name': ReconfigVM_Task, 'duration_secs': 0.227657} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.984633] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Reconfigured VM instance instance-0000005e to detach disk 2000 {{(pid=63531) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1030.985422] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c081eeb1-6f07-4a8e-84b5-a08cb9f4d5a6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.989437] env[63531]: DEBUG nova.scheduler.client.report [None req-f1609b7f-e5ee-4b90-9864-431682b40af5 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1031.016772] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Reconfiguring VM instance instance-0000005e to attach disk [datastore2] 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7/5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1031.018108] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-df58486a-739d-4740-ab7c-7ccf0431b934 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.037459] env[63531]: DEBUG oslo_concurrency.lockutils [None req-31d8ac8d-3854-40f0-8430-70551d6a96ed tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquiring lock "4757625a-8398-4b2d-b6fe-40eea8913068" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.038144] env[63531]: DEBUG oslo_concurrency.lockutils [None req-31d8ac8d-3854-40f0-8430-70551d6a96ed tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "4757625a-8398-4b2d-b6fe-40eea8913068" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.040856] env[63531]: DEBUG oslo_vmware.api [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 1031.040856] env[63531]: value = "task-1118468" [ 1031.040856] env[63531]: _type = "Task" [ 1031.040856] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.051732] env[63531]: DEBUG oslo_vmware.api [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118468, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.132202] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.216211] env[63531]: DEBUG nova.objects.instance [None req-c12caf0f-2eba-4205-bb3b-92d592445017 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Lazy-loading 'flavor' on Instance uuid ccf00b09-29fe-4a76-a8af-97774acd77de {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1031.258952] env[63531]: DEBUG nova.network.neutron [req-2cae8d63-c355-4bd4-ac4a-0c5bb158463a req-57d2e6c2-2776-499b-b6bf-fd5e430702d7 service nova] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Updated VIF entry in instance network info cache for port b731766f-738c-4d77-a00e-98ed6c683e55. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1031.258952] env[63531]: DEBUG nova.network.neutron [req-2cae8d63-c355-4bd4-ac4a-0c5bb158463a req-57d2e6c2-2776-499b-b6bf-fd5e430702d7 service nova] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Updating instance_info_cache with network_info: [{"id": "b731766f-738c-4d77-a00e-98ed6c683e55", "address": "fa:16:3e:d5:26:02", "network": {"id": "0667c176-cfdf-4e43-b152-e00bdc224448", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-243715985-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.217", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "308eabafea9c4188a58a0f1c22074d2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8f580e6-1d86-41ee-9ebe-c531cb9299c6", "external-id": "nsx-vlan-transportzone-150", "segmentation_id": 150, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb731766f-73", "ovs_interfaceid": "b731766f-738c-4d77-a00e-98ed6c683e55", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1031.311978] env[63531]: DEBUG nova.compute.manager [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 97d3f108-e299-4ef1-84ca-532cd64905d5] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1031.496943] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f1609b7f-e5ee-4b90-9864-431682b40af5 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.893s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.500387] env[63531]: DEBUG oslo_concurrency.lockutils [None req-89655885-53ad-493a-9114-f2b1c23800d2 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.864s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.500808] env[63531]: DEBUG nova.objects.instance [None req-89655885-53ad-493a-9114-f2b1c23800d2 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Lazy-loading 'resources' on Instance uuid afa5f2c8-e519-4eb7-8e0e-ada9efb630e7 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1031.522256] env[63531]: INFO nova.scheduler.client.report [None req-f1609b7f-e5ee-4b90-9864-431682b40af5 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Deleted allocations for instance ba97ce22-ea4f-430b-a4c6-08204a568436 [ 1031.543020] env[63531]: DEBUG nova.compute.utils [None req-31d8ac8d-3854-40f0-8430-70551d6a96ed tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1031.554078] env[63531]: DEBUG oslo_concurrency.lockutils [None req-31d8ac8d-3854-40f0-8430-70551d6a96ed tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "4757625a-8398-4b2d-b6fe-40eea8913068" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 0.516s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.561693] env[63531]: DEBUG oslo_vmware.api [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118468, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.761566] env[63531]: DEBUG oslo_concurrency.lockutils [req-2cae8d63-c355-4bd4-ac4a-0c5bb158463a req-57d2e6c2-2776-499b-b6bf-fd5e430702d7 service nova] Releasing lock "refresh_cache-96c2910b-1e25-4053-9ad9-c155402ef0ac" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1031.831948] env[63531]: DEBUG oslo_concurrency.lockutils [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.038475] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f1609b7f-e5ee-4b90-9864-431682b40af5 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Lock "ba97ce22-ea4f-430b-a4c6-08204a568436" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.348s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.055846] env[63531]: DEBUG oslo_vmware.api [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118468, 'name': ReconfigVM_Task, 'duration_secs': 0.53096} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.056829] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Reconfigured VM instance instance-0000005e to attach disk [datastore2] 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7/5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1032.057185] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Updating instance '5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7' progress to 50 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1032.224499] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c12caf0f-2eba-4205-bb3b-92d592445017 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Lock "ccf00b09-29fe-4a76-a8af-97774acd77de" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.282s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.263931] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0019760b-89a9-4ce4-a3c8-0692ad6133d4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.272186] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9baaccf-a97c-4d2c-8deb-3838fc8d9a11 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.305096] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee04ae27-79b1-4804-b322-e56db0e495b3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.313772] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11962479-dcf5-4f12-83b0-2939bc0fba8e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.327550] env[63531]: DEBUG nova.compute.provider_tree [None req-89655885-53ad-493a-9114-f2b1c23800d2 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1032.564145] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93aca94f-8292-4936-990a-93e1e92b4e65 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.584504] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e6bb3d3-961b-4a8c-bb6a-b2ada9a52b11 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.602776] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Updating instance '5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7' progress to 67 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1032.622570] env[63531]: DEBUG oslo_concurrency.lockutils [None req-31d8ac8d-3854-40f0-8430-70551d6a96ed tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquiring lock "4757625a-8398-4b2d-b6fe-40eea8913068" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.622849] env[63531]: DEBUG oslo_concurrency.lockutils [None req-31d8ac8d-3854-40f0-8430-70551d6a96ed tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "4757625a-8398-4b2d-b6fe-40eea8913068" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.623115] env[63531]: INFO nova.compute.manager [None req-31d8ac8d-3854-40f0-8430-70551d6a96ed tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Attaching volume 8537bea9-d5a3-4b76-84eb-8774f4539325 to /dev/sdb [ 1032.659044] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1da9d7e3-6a50-4719-9c08-aeccff052a34 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.665968] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f506060-1302-4096-8c3b-13cdaaf138d7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.672492] env[63531]: DEBUG oslo_concurrency.lockutils [None req-df4e3d53-9f91-4cef-a886-ae965b0f89e8 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquiring lock "ccf00b09-29fe-4a76-a8af-97774acd77de" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.672748] env[63531]: DEBUG oslo_concurrency.lockutils [None req-df4e3d53-9f91-4cef-a886-ae965b0f89e8 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Lock "ccf00b09-29fe-4a76-a8af-97774acd77de" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.672971] env[63531]: DEBUG oslo_concurrency.lockutils [None req-df4e3d53-9f91-4cef-a886-ae965b0f89e8 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquiring lock "ccf00b09-29fe-4a76-a8af-97774acd77de-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.673196] env[63531]: DEBUG oslo_concurrency.lockutils [None req-df4e3d53-9f91-4cef-a886-ae965b0f89e8 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Lock "ccf00b09-29fe-4a76-a8af-97774acd77de-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.673374] env[63531]: DEBUG oslo_concurrency.lockutils [None req-df4e3d53-9f91-4cef-a886-ae965b0f89e8 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Lock "ccf00b09-29fe-4a76-a8af-97774acd77de-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.675580] env[63531]: INFO nova.compute.manager [None req-df4e3d53-9f91-4cef-a886-ae965b0f89e8 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Terminating instance [ 1032.680885] env[63531]: DEBUG nova.virt.block_device [None req-31d8ac8d-3854-40f0-8430-70551d6a96ed tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Updating existing volume attachment record: 73ed93b0-cd3e-43a0-970b-e68de3f4ed40 {{(pid=63531) _volume_attach /opt/stack/nova/nova/virt/block_device.py:679}} [ 1032.683604] env[63531]: DEBUG nova.compute.manager [None req-df4e3d53-9f91-4cef-a886-ae965b0f89e8 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1032.683802] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-df4e3d53-9f91-4cef-a886-ae965b0f89e8 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1032.684845] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7df872e9-fccd-48c6-a142-b8cf69e6bdca {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.692043] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-df4e3d53-9f91-4cef-a886-ae965b0f89e8 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1032.692275] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c2ea5ad9-28f8-4286-b148-c0e2d4bf9d69 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.698529] env[63531]: DEBUG oslo_vmware.api [None req-df4e3d53-9f91-4cef-a886-ae965b0f89e8 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 1032.698529] env[63531]: value = "task-1118469" [ 1032.698529] env[63531]: _type = "Task" [ 1032.698529] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.707395] env[63531]: DEBUG oslo_vmware.api [None req-df4e3d53-9f91-4cef-a886-ae965b0f89e8 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118469, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.721637] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7881f65b-c8d5-4daf-91f4-da68e00cf697 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "interface-e742d3af-5a45-4274-adfb-c6138bf84d6a-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.722248] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7881f65b-c8d5-4daf-91f4-da68e00cf697 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "interface-e742d3af-5a45-4274-adfb-c6138bf84d6a-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.723926] env[63531]: DEBUG nova.objects.instance [None req-7881f65b-c8d5-4daf-91f4-da68e00cf697 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lazy-loading 'flavor' on Instance uuid e742d3af-5a45-4274-adfb-c6138bf84d6a {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1032.830598] env[63531]: DEBUG nova.scheduler.client.report [None req-89655885-53ad-493a-9114-f2b1c23800d2 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1033.171882] env[63531]: DEBUG nova.network.neutron [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Port 91a7c60f-cbcb-4fe3-9bd0-69337882b7ce binding to destination host cpu-1 is already ACTIVE {{(pid=63531) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1033.208701] env[63531]: DEBUG oslo_vmware.api [None req-df4e3d53-9f91-4cef-a886-ae965b0f89e8 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118469, 'name': PowerOffVM_Task, 'duration_secs': 0.251173} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.208968] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-df4e3d53-9f91-4cef-a886-ae965b0f89e8 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1033.209237] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-df4e3d53-9f91-4cef-a886-ae965b0f89e8 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1033.209508] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bd41d3d3-883a-4050-905a-20408d06e87f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.227568] env[63531]: DEBUG nova.objects.instance [None req-7881f65b-c8d5-4daf-91f4-da68e00cf697 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lazy-loading 'pci_requests' on Instance uuid e742d3af-5a45-4274-adfb-c6138bf84d6a {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1033.338343] env[63531]: DEBUG oslo_concurrency.lockutils [None req-89655885-53ad-493a-9114-f2b1c23800d2 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.838s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.342019] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.209s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.342915] env[63531]: INFO nova.compute.claims [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1033.362973] env[63531]: INFO nova.scheduler.client.report [None req-89655885-53ad-493a-9114-f2b1c23800d2 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Deleted allocations for instance afa5f2c8-e519-4eb7-8e0e-ada9efb630e7 [ 1033.731017] env[63531]: DEBUG nova.objects.base [None req-7881f65b-c8d5-4daf-91f4-da68e00cf697 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=63531) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1033.731576] env[63531]: DEBUG nova.network.neutron [None req-7881f65b-c8d5-4daf-91f4-da68e00cf697 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1033.871015] env[63531]: DEBUG oslo_concurrency.lockutils [None req-89655885-53ad-493a-9114-f2b1c23800d2 tempest-MultipleCreateTestJSON-163192523 tempest-MultipleCreateTestJSON-163192523-project-member] Lock "afa5f2c8-e519-4eb7-8e0e-ada9efb630e7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.205s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.102491] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7881f65b-c8d5-4daf-91f4-da68e00cf697 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "interface-e742d3af-5a45-4274-adfb-c6138bf84d6a-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.380s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.195598] env[63531]: DEBUG oslo_concurrency.lockutils [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.195871] env[63531]: DEBUG oslo_concurrency.lockutils [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.196059] env[63531]: DEBUG oslo_concurrency.lockutils [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.549725] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ee11d3f-1de6-4975-875d-d81fe3f5474f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.557654] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61bf8106-1c04-41ea-a6b4-d16f2aae0835 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.588920] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3330ba5-8b72-450b-b410-36ab73446d30 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.596723] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23d5af8f-3c5b-465a-93c8-d83c06d07398 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.612339] env[63531]: DEBUG nova.compute.provider_tree [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1034.678491] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-df4e3d53-9f91-4cef-a886-ae965b0f89e8 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1034.678925] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-df4e3d53-9f91-4cef-a886-ae965b0f89e8 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1034.679221] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-df4e3d53-9f91-4cef-a886-ae965b0f89e8 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Deleting the datastore file [datastore1] ccf00b09-29fe-4a76-a8af-97774acd77de {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1034.679622] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6ced2f25-c20a-4103-af02-d0b474e6aef3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.686876] env[63531]: DEBUG oslo_vmware.api [None req-df4e3d53-9f91-4cef-a886-ae965b0f89e8 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 1034.686876] env[63531]: value = "task-1118472" [ 1034.686876] env[63531]: _type = "Task" [ 1034.686876] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.695158] env[63531]: DEBUG oslo_vmware.api [None req-df4e3d53-9f91-4cef-a886-ae965b0f89e8 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118472, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.119084] env[63531]: DEBUG nova.scheduler.client.report [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1035.197467] env[63531]: DEBUG oslo_vmware.api [None req-df4e3d53-9f91-4cef-a886-ae965b0f89e8 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118472, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.39557} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.200434] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-df4e3d53-9f91-4cef-a886-ae965b0f89e8 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1035.200646] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-df4e3d53-9f91-4cef-a886-ae965b0f89e8 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1035.200832] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-df4e3d53-9f91-4cef-a886-ae965b0f89e8 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1035.201023] env[63531]: INFO nova.compute.manager [None req-df4e3d53-9f91-4cef-a886-ae965b0f89e8 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Took 2.52 seconds to destroy the instance on the hypervisor. [ 1035.201279] env[63531]: DEBUG oslo.service.loopingcall [None req-df4e3d53-9f91-4cef-a886-ae965b0f89e8 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1035.201495] env[63531]: DEBUG nova.compute.manager [-] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1035.201604] env[63531]: DEBUG nova.network.neutron [-] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1035.278530] env[63531]: DEBUG oslo_concurrency.lockutils [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "refresh_cache-5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1035.278893] env[63531]: DEBUG oslo_concurrency.lockutils [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquired lock "refresh_cache-5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.278893] env[63531]: DEBUG nova.network.neutron [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1035.624951] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.284s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.625556] env[63531]: DEBUG nova.compute.manager [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1035.629181] env[63531]: DEBUG oslo_concurrency.lockutils [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.797s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.633127] env[63531]: INFO nova.compute.claims [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 97d3f108-e299-4ef1-84ca-532cd64905d5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1035.771018] env[63531]: DEBUG nova.compute.manager [req-92aa1502-58d5-4e46-b7ed-8d047d8c1aa1 req-4cc2f27c-9e0e-4bf5-988e-65b9e46b40c1 service nova] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Received event network-vif-deleted-7433ce46-ff5a-4e78-bcc9-9dcf2264a01b {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1035.771018] env[63531]: INFO nova.compute.manager [req-92aa1502-58d5-4e46-b7ed-8d047d8c1aa1 req-4cc2f27c-9e0e-4bf5-988e-65b9e46b40c1 service nova] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Neutron deleted interface 7433ce46-ff5a-4e78-bcc9-9dcf2264a01b; detaching it from the instance and deleting it from the info cache [ 1035.771018] env[63531]: DEBUG nova.network.neutron [req-92aa1502-58d5-4e46-b7ed-8d047d8c1aa1 req-4cc2f27c-9e0e-4bf5-988e-65b9e46b40c1 service nova] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1035.771394] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3afa2054-43b4-4470-afb8-c9c445c20c12 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "interface-e742d3af-5a45-4274-adfb-c6138bf84d6a-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.771897] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3afa2054-43b4-4470-afb8-c9c445c20c12 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "interface-e742d3af-5a45-4274-adfb-c6138bf84d6a-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.771897] env[63531]: DEBUG nova.objects.instance [None req-3afa2054-43b4-4470-afb8-c9c445c20c12 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lazy-loading 'flavor' on Instance uuid e742d3af-5a45-4274-adfb-c6138bf84d6a {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1036.041984] env[63531]: DEBUG nova.network.neutron [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Updating instance_info_cache with network_info: [{"id": "91a7c60f-cbcb-4fe3-9bd0-69337882b7ce", "address": "fa:16:3e:46:61:c6", "network": {"id": "bf99b8cb-9efb-4a74-9db1-0f5f0875f26a", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-998205595-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2e99349303e413b9f7896d449cb7dff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc6e6fe1-c4f5-4389-a49f-0978060eebb4", "external-id": "nsx-vlan-transportzone-829", "segmentation_id": 829, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91a7c60f-cb", "ovs_interfaceid": "91a7c60f-cbcb-4fe3-9bd0-69337882b7ce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1036.136494] env[63531]: DEBUG nova.compute.utils [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1036.140011] env[63531]: DEBUG nova.compute.manager [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1036.140011] env[63531]: DEBUG nova.network.neutron [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1036.180967] env[63531]: DEBUG nova.policy [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0a501b2c64c04dcda3596c8a1bd3fa0d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'abdae960d43b47bc95ab41f265d2c85b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 1036.245706] env[63531]: DEBUG nova.network.neutron [-] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1036.275140] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b7b77946-7342-4d93-bb81-60a214de7e1c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.289451] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dae74135-8296-4fd1-9943-69d0eca240c4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.322888] env[63531]: DEBUG nova.compute.manager [req-92aa1502-58d5-4e46-b7ed-8d047d8c1aa1 req-4cc2f27c-9e0e-4bf5-988e-65b9e46b40c1 service nova] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Detach interface failed, port_id=7433ce46-ff5a-4e78-bcc9-9dcf2264a01b, reason: Instance ccf00b09-29fe-4a76-a8af-97774acd77de could not be found. {{(pid=63531) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1036.385127] env[63531]: DEBUG nova.objects.instance [None req-3afa2054-43b4-4470-afb8-c9c445c20c12 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lazy-loading 'pci_requests' on Instance uuid e742d3af-5a45-4274-adfb-c6138bf84d6a {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1036.451577] env[63531]: DEBUG nova.network.neutron [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] Successfully created port: b8e8e4e6-84b5-46b4-9b14-65eb203d51cb {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1036.546619] env[63531]: DEBUG oslo_concurrency.lockutils [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Releasing lock "refresh_cache-5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1036.641123] env[63531]: DEBUG nova.compute.manager [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1036.750546] env[63531]: INFO nova.compute.manager [-] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Took 1.55 seconds to deallocate network for instance. [ 1036.882399] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f610bc2-e54e-4715-9a42-7700d39be7d8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.888067] env[63531]: DEBUG nova.objects.base [None req-3afa2054-43b4-4470-afb8-c9c445c20c12 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=63531) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1036.888360] env[63531]: DEBUG nova.network.neutron [None req-3afa2054-43b4-4470-afb8-c9c445c20c12 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1036.893587] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6278ceb-0f7e-4d8e-a92c-e6b536399c00 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.925546] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-256f5d6d-527f-4e59-b624-eeb185d13611 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.937080] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf2a999f-aa2b-4a6e-bee7-190a10e86f51 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.951543] env[63531]: DEBUG nova.compute.provider_tree [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1036.957107] env[63531]: DEBUG nova.policy [None req-3afa2054-43b4-4470-afb8-c9c445c20c12 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '36cb2f0de54f417384fa1fb2210de4e4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8ba8b284854242c392aec5326e996239', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 1036.966210] env[63531]: DEBUG oslo_vmware.rw_handles [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e47cd7-a2be-aef4-7628-cbce47f138d1/disk-0.vmdk. {{(pid=63531) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1036.969205] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1b6d5ad-b155-45b3-b945-30c116cf3566 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.973648] env[63531]: DEBUG oslo_vmware.rw_handles [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e47cd7-a2be-aef4-7628-cbce47f138d1/disk-0.vmdk is in state: ready. {{(pid=63531) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1036.973885] env[63531]: ERROR oslo_vmware.rw_handles [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e47cd7-a2be-aef4-7628-cbce47f138d1/disk-0.vmdk due to incomplete transfer. [ 1036.974147] env[63531]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-df2897bf-6c31-48c6-b35c-cbaa15fd89f2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.982925] env[63531]: DEBUG oslo_vmware.rw_handles [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e47cd7-a2be-aef4-7628-cbce47f138d1/disk-0.vmdk. {{(pid=63531) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1036.982925] env[63531]: DEBUG nova.virt.vmwareapi.images [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Uploaded image 87aadec8-cd66-47c5-a3d7-3ae819edebf8 to the Glance image server {{(pid=63531) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1036.985065] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Destroying the VM {{(pid=63531) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1036.985444] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-75fa2e51-ad4a-4bec-a7c7-528398e43ff7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.992433] env[63531]: DEBUG oslo_vmware.api [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 1036.992433] env[63531]: value = "task-1118474" [ 1036.992433] env[63531]: _type = "Task" [ 1036.992433] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.001473] env[63531]: DEBUG oslo_vmware.api [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118474, 'name': Destroy_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.071576] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-920c3a8e-c304-42ab-99f2-508ad2f230b0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.092523] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e068547e-9ef4-409a-80b7-3bfe1223d8a8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.099799] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Updating instance '5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7' progress to 83 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1037.240657] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-31d8ac8d-3854-40f0-8430-70551d6a96ed tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Volume attach. Driver type: vmdk {{(pid=63531) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1037.240922] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-31d8ac8d-3854-40f0-8430-70551d6a96ed tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244788', 'volume_id': '8537bea9-d5a3-4b76-84eb-8774f4539325', 'name': 'volume-8537bea9-d5a3-4b76-84eb-8774f4539325', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4757625a-8398-4b2d-b6fe-40eea8913068', 'attached_at': '', 'detached_at': '', 'volume_id': '8537bea9-d5a3-4b76-84eb-8774f4539325', 'serial': '8537bea9-d5a3-4b76-84eb-8774f4539325'} {{(pid=63531) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1037.241828] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fbb66bc-6014-45bb-815d-f6dde1036a02 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.258776] env[63531]: DEBUG nova.network.neutron [None req-3afa2054-43b4-4470-afb8-c9c445c20c12 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Successfully created port: 5f33c6bf-e31f-422e-8611-c4efb3719681 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1037.261175] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02ef2687-2104-4907-ba6d-ba716bed1c18 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.264214] env[63531]: DEBUG oslo_concurrency.lockutils [None req-df4e3d53-9f91-4cef-a886-ae965b0f89e8 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.287507] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-31d8ac8d-3854-40f0-8430-70551d6a96ed tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Reconfiguring VM instance instance-0000005a to attach disk [datastore2] volume-8537bea9-d5a3-4b76-84eb-8774f4539325/volume-8537bea9-d5a3-4b76-84eb-8774f4539325.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1037.287842] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-32d0645a-129f-4080-bcdb-e0bfe7e69757 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.305959] env[63531]: DEBUG oslo_vmware.api [None req-31d8ac8d-3854-40f0-8430-70551d6a96ed tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for the task: (returnval){ [ 1037.305959] env[63531]: value = "task-1118475" [ 1037.305959] env[63531]: _type = "Task" [ 1037.305959] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.314724] env[63531]: DEBUG oslo_vmware.api [None req-31d8ac8d-3854-40f0-8430-70551d6a96ed tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118475, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.459406] env[63531]: DEBUG nova.scheduler.client.report [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1037.501916] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Acquiring lock "3e7b8a01-d930-43cf-9432-0dabc68da20e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.502227] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Lock "3e7b8a01-d930-43cf-9432-0dabc68da20e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.510395] env[63531]: DEBUG oslo_vmware.api [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118474, 'name': Destroy_Task, 'duration_secs': 0.378487} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.510716] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Destroyed the VM [ 1037.510943] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Deleting Snapshot of the VM instance {{(pid=63531) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1037.511232] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-d8b47c62-d2e4-4dfd-8342-61947d580997 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.517978] env[63531]: DEBUG oslo_vmware.api [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 1037.517978] env[63531]: value = "task-1118476" [ 1037.517978] env[63531]: _type = "Task" [ 1037.517978] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.526334] env[63531]: DEBUG oslo_vmware.api [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118476, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.606331] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1037.606728] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5149d50c-e551-4e46-bdf7-5c1bf1808364 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.614529] env[63531]: DEBUG oslo_vmware.api [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 1037.614529] env[63531]: value = "task-1118477" [ 1037.614529] env[63531]: _type = "Task" [ 1037.614529] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.622631] env[63531]: DEBUG oslo_vmware.api [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118477, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.652746] env[63531]: DEBUG nova.compute.manager [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1037.676485] env[63531]: DEBUG nova.virt.hardware [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1037.676756] env[63531]: DEBUG nova.virt.hardware [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1037.676917] env[63531]: DEBUG nova.virt.hardware [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1037.677123] env[63531]: DEBUG nova.virt.hardware [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1037.677282] env[63531]: DEBUG nova.virt.hardware [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1037.677434] env[63531]: DEBUG nova.virt.hardware [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1037.677680] env[63531]: DEBUG nova.virt.hardware [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1037.677867] env[63531]: DEBUG nova.virt.hardware [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1037.678056] env[63531]: DEBUG nova.virt.hardware [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1037.678259] env[63531]: DEBUG nova.virt.hardware [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1037.678448] env[63531]: DEBUG nova.virt.hardware [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1037.679557] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9f1fa7e-6671-488a-a68b-b2e32b2041c4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.686952] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf067b22-046a-4942-8f25-048466df2d63 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.815855] env[63531]: DEBUG oslo_vmware.api [None req-31d8ac8d-3854-40f0-8430-70551d6a96ed tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118475, 'name': ReconfigVM_Task, 'duration_secs': 0.352907} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.816350] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-31d8ac8d-3854-40f0-8430-70551d6a96ed tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Reconfigured VM instance instance-0000005a to attach disk [datastore2] volume-8537bea9-d5a3-4b76-84eb-8774f4539325/volume-8537bea9-d5a3-4b76-84eb-8774f4539325.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1037.821376] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-244ae876-7af3-4c42-9926-81c3aec92660 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.837024] env[63531]: DEBUG oslo_vmware.api [None req-31d8ac8d-3854-40f0-8430-70551d6a96ed tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for the task: (returnval){ [ 1037.837024] env[63531]: value = "task-1118478" [ 1037.837024] env[63531]: _type = "Task" [ 1037.837024] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.846295] env[63531]: DEBUG oslo_vmware.api [None req-31d8ac8d-3854-40f0-8430-70551d6a96ed tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118478, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.909788] env[63531]: DEBUG nova.compute.manager [req-cb1de809-cc72-46c3-9011-36a4af3c09e2 req-1daf9c8b-587b-453f-82d4-686335b101aa service nova] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] Received event network-vif-plugged-b8e8e4e6-84b5-46b4-9b14-65eb203d51cb {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1037.910043] env[63531]: DEBUG oslo_concurrency.lockutils [req-cb1de809-cc72-46c3-9011-36a4af3c09e2 req-1daf9c8b-587b-453f-82d4-686335b101aa service nova] Acquiring lock "c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.910260] env[63531]: DEBUG oslo_concurrency.lockutils [req-cb1de809-cc72-46c3-9011-36a4af3c09e2 req-1daf9c8b-587b-453f-82d4-686335b101aa service nova] Lock "c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.910449] env[63531]: DEBUG oslo_concurrency.lockutils [req-cb1de809-cc72-46c3-9011-36a4af3c09e2 req-1daf9c8b-587b-453f-82d4-686335b101aa service nova] Lock "c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.910644] env[63531]: DEBUG nova.compute.manager [req-cb1de809-cc72-46c3-9011-36a4af3c09e2 req-1daf9c8b-587b-453f-82d4-686335b101aa service nova] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] No waiting events found dispatching network-vif-plugged-b8e8e4e6-84b5-46b4-9b14-65eb203d51cb {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1037.910817] env[63531]: WARNING nova.compute.manager [req-cb1de809-cc72-46c3-9011-36a4af3c09e2 req-1daf9c8b-587b-453f-82d4-686335b101aa service nova] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] Received unexpected event network-vif-plugged-b8e8e4e6-84b5-46b4-9b14-65eb203d51cb for instance with vm_state building and task_state spawning. [ 1037.967319] env[63531]: DEBUG oslo_concurrency.lockutils [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.338s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.968383] env[63531]: DEBUG nova.compute.manager [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 97d3f108-e299-4ef1-84ca-532cd64905d5] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1037.971462] env[63531]: DEBUG oslo_concurrency.lockutils [None req-df4e3d53-9f91-4cef-a886-ae965b0f89e8 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.707s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.971725] env[63531]: DEBUG nova.objects.instance [None req-df4e3d53-9f91-4cef-a886-ae965b0f89e8 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Lazy-loading 'resources' on Instance uuid ccf00b09-29fe-4a76-a8af-97774acd77de {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1038.007495] env[63531]: DEBUG nova.compute.manager [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] [instance: 3e7b8a01-d930-43cf-9432-0dabc68da20e] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1038.030695] env[63531]: DEBUG oslo_vmware.api [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118476, 'name': RemoveSnapshot_Task, 'duration_secs': 0.398609} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.034703] env[63531]: DEBUG nova.network.neutron [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] Successfully updated port: b8e8e4e6-84b5-46b4-9b14-65eb203d51cb {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1038.036776] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Deleted Snapshot of the VM instance {{(pid=63531) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1038.037210] env[63531]: DEBUG nova.compute.manager [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1038.038852] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9361bf0c-c8a4-47d3-9e57-41131888d89d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.125156] env[63531]: DEBUG oslo_vmware.api [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118477, 'name': PowerOnVM_Task, 'duration_secs': 0.466418} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.125531] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1038.125733] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-67beaa6b-3fbf-4291-a632-55c106df4795 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Updating instance '5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7' progress to 100 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1038.347195] env[63531]: DEBUG oslo_vmware.api [None req-31d8ac8d-3854-40f0-8430-70551d6a96ed tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118478, 'name': ReconfigVM_Task, 'duration_secs': 0.194103} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.347533] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-31d8ac8d-3854-40f0-8430-70551d6a96ed tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244788', 'volume_id': '8537bea9-d5a3-4b76-84eb-8774f4539325', 'name': 'volume-8537bea9-d5a3-4b76-84eb-8774f4539325', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4757625a-8398-4b2d-b6fe-40eea8913068', 'attached_at': '', 'detached_at': '', 'volume_id': '8537bea9-d5a3-4b76-84eb-8774f4539325', 'serial': '8537bea9-d5a3-4b76-84eb-8774f4539325'} {{(pid=63531) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1038.475092] env[63531]: DEBUG nova.compute.utils [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1038.480788] env[63531]: DEBUG nova.compute.manager [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 97d3f108-e299-4ef1-84ca-532cd64905d5] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1038.480788] env[63531]: DEBUG nova.network.neutron [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 97d3f108-e299-4ef1-84ca-532cd64905d5] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1038.528618] env[63531]: DEBUG nova.policy [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7140e12106504b3e95ff49786361771b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '63900312a30b4942854f552758c039cf', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 1038.530762] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.539095] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquiring lock "refresh_cache-c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1038.539253] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquired lock "refresh_cache-c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1038.539473] env[63531]: DEBUG nova.network.neutron [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1038.556553] env[63531]: INFO nova.compute.manager [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Shelve offloading [ 1038.557708] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1038.557957] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-92808238-dc0f-4816-9269-f2c393520254 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.565139] env[63531]: DEBUG oslo_vmware.api [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 1038.565139] env[63531]: value = "task-1118479" [ 1038.565139] env[63531]: _type = "Task" [ 1038.565139] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.583843] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] VM already powered off {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1038.584009] env[63531]: DEBUG nova.compute.manager [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1038.584876] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e360826-2224-459e-a917-6daa56d9e0d5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.590696] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "refresh_cache-3fe7a7a8-fc43-4560-b101-227907c91fd0" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1038.590851] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquired lock "refresh_cache-3fe7a7a8-fc43-4560-b101-227907c91fd0" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1038.591125] env[63531]: DEBUG nova.network.neutron [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1038.726271] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-653fd7cc-d7bd-4584-9f99-a5f03d6cfe10 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.739838] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85b43345-aa93-40f0-9cc9-84a5c0d0a776 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.778698] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad5e57dd-91ae-4e72-96ac-4fbb08886e00 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.786880] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-965926ed-f8e6-4ad8-803b-77db3daca688 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.800425] env[63531]: DEBUG nova.compute.provider_tree [None req-df4e3d53-9f91-4cef-a886-ae965b0f89e8 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1038.802272] env[63531]: DEBUG nova.network.neutron [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 97d3f108-e299-4ef1-84ca-532cd64905d5] Successfully created port: b9f690d1-15fa-46cb-a818-c858598dd6f9 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1038.868506] env[63531]: DEBUG nova.network.neutron [None req-3afa2054-43b4-4470-afb8-c9c445c20c12 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Successfully updated port: 5f33c6bf-e31f-422e-8611-c4efb3719681 {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1038.886356] env[63531]: DEBUG nova.compute.manager [req-8adf0946-5318-48fb-88c2-494fb03f0799 req-2daafe88-bc18-4cea-be6e-ef4b94cc4885 service nova] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Received event network-vif-plugged-5f33c6bf-e31f-422e-8611-c4efb3719681 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1038.887184] env[63531]: DEBUG oslo_concurrency.lockutils [req-8adf0946-5318-48fb-88c2-494fb03f0799 req-2daafe88-bc18-4cea-be6e-ef4b94cc4885 service nova] Acquiring lock "e742d3af-5a45-4274-adfb-c6138bf84d6a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.887184] env[63531]: DEBUG oslo_concurrency.lockutils [req-8adf0946-5318-48fb-88c2-494fb03f0799 req-2daafe88-bc18-4cea-be6e-ef4b94cc4885 service nova] Lock "e742d3af-5a45-4274-adfb-c6138bf84d6a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.887184] env[63531]: DEBUG oslo_concurrency.lockutils [req-8adf0946-5318-48fb-88c2-494fb03f0799 req-2daafe88-bc18-4cea-be6e-ef4b94cc4885 service nova] Lock "e742d3af-5a45-4274-adfb-c6138bf84d6a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.887422] env[63531]: DEBUG nova.compute.manager [req-8adf0946-5318-48fb-88c2-494fb03f0799 req-2daafe88-bc18-4cea-be6e-ef4b94cc4885 service nova] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] No waiting events found dispatching network-vif-plugged-5f33c6bf-e31f-422e-8611-c4efb3719681 {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1038.887603] env[63531]: WARNING nova.compute.manager [req-8adf0946-5318-48fb-88c2-494fb03f0799 req-2daafe88-bc18-4cea-be6e-ef4b94cc4885 service nova] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Received unexpected event network-vif-plugged-5f33c6bf-e31f-422e-8611-c4efb3719681 for instance with vm_state active and task_state None. [ 1038.979901] env[63531]: DEBUG nova.compute.manager [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 97d3f108-e299-4ef1-84ca-532cd64905d5] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1039.116256] env[63531]: DEBUG nova.network.neutron [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1039.305123] env[63531]: DEBUG nova.scheduler.client.report [None req-df4e3d53-9f91-4cef-a886-ae965b0f89e8 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1039.376035] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3afa2054-43b4-4470-afb8-c9c445c20c12 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "refresh_cache-e742d3af-5a45-4274-adfb-c6138bf84d6a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1039.376454] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3afa2054-43b4-4470-afb8-c9c445c20c12 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquired lock "refresh_cache-e742d3af-5a45-4274-adfb-c6138bf84d6a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1039.376615] env[63531]: DEBUG nova.network.neutron [None req-3afa2054-43b4-4470-afb8-c9c445c20c12 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1039.405047] env[63531]: DEBUG nova.objects.instance [None req-31d8ac8d-3854-40f0-8430-70551d6a96ed tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lazy-loading 'flavor' on Instance uuid 4757625a-8398-4b2d-b6fe-40eea8913068 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1039.447541] env[63531]: DEBUG nova.network.neutron [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] Updating instance_info_cache with network_info: [{"id": "b8e8e4e6-84b5-46b4-9b14-65eb203d51cb", "address": "fa:16:3e:99:13:e7", "network": {"id": "eac8786d-bc67-4916-8483-b53d9a25efb8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1604989690-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "abdae960d43b47bc95ab41f265d2c85b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db068f71-08cc-42d4-8ab6-17134c1585e5", "external-id": "nsx-vlan-transportzone-721", "segmentation_id": 721, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb8e8e4e6-84", "ovs_interfaceid": "b8e8e4e6-84b5-46b4-9b14-65eb203d51cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1039.489387] env[63531]: DEBUG nova.network.neutron [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Updating instance_info_cache with network_info: [{"id": "03e99378-6a13-403e-9cb4-856abf141914", "address": "fa:16:3e:3d:57:da", "network": {"id": "ea2d47d3-a502-4afd-9bc5-ee4a37017901", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1361282648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94a95667c00947ea9183a6307c569c90", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6c45fd-e930-495a-9cb7-df84eda443b1", "external-id": "nsx-vlan-transportzone-407", "segmentation_id": 407, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03e99378-6a", "ovs_interfaceid": "03e99378-6a13-403e-9cb4-856abf141914", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1039.809998] env[63531]: DEBUG oslo_concurrency.lockutils [None req-df4e3d53-9f91-4cef-a886-ae965b0f89e8 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.838s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.812704] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.282s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.814219] env[63531]: INFO nova.compute.claims [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] [instance: 3e7b8a01-d930-43cf-9432-0dabc68da20e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1039.835726] env[63531]: INFO nova.scheduler.client.report [None req-df4e3d53-9f91-4cef-a886-ae965b0f89e8 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Deleted allocations for instance ccf00b09-29fe-4a76-a8af-97774acd77de [ 1039.909798] env[63531]: DEBUG oslo_concurrency.lockutils [None req-31d8ac8d-3854-40f0-8430-70551d6a96ed tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "4757625a-8398-4b2d-b6fe-40eea8913068" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.287s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.924896] env[63531]: WARNING nova.network.neutron [None req-3afa2054-43b4-4470-afb8-c9c445c20c12 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] caf86428-59a5-4e50-9135-2ab8473af64d already exists in list: networks containing: ['caf86428-59a5-4e50-9135-2ab8473af64d']. ignoring it [ 1039.954833] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Releasing lock "refresh_cache-c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1039.955176] env[63531]: DEBUG nova.compute.manager [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] Instance network_info: |[{"id": "b8e8e4e6-84b5-46b4-9b14-65eb203d51cb", "address": "fa:16:3e:99:13:e7", "network": {"id": "eac8786d-bc67-4916-8483-b53d9a25efb8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1604989690-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "abdae960d43b47bc95ab41f265d2c85b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db068f71-08cc-42d4-8ab6-17134c1585e5", "external-id": "nsx-vlan-transportzone-721", "segmentation_id": 721, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb8e8e4e6-84", "ovs_interfaceid": "b8e8e4e6-84b5-46b4-9b14-65eb203d51cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1039.955638] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:99:13:e7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'db068f71-08cc-42d4-8ab6-17134c1585e5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b8e8e4e6-84b5-46b4-9b14-65eb203d51cb', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1039.963086] env[63531]: DEBUG oslo.service.loopingcall [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1039.963311] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1039.963536] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1b942c9a-3860-4b96-96d7-fd75ebfce77b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.985487] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1039.985487] env[63531]: value = "task-1118480" [ 1039.985487] env[63531]: _type = "Task" [ 1039.985487] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.989633] env[63531]: DEBUG nova.compute.manager [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 97d3f108-e299-4ef1-84ca-532cd64905d5] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1039.993198] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Releasing lock "refresh_cache-3fe7a7a8-fc43-4560-b101-227907c91fd0" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1039.999463] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118480, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.015776] env[63531]: DEBUG nova.virt.hardware [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1040.016026] env[63531]: DEBUG nova.virt.hardware [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1040.016188] env[63531]: DEBUG nova.virt.hardware [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1040.016395] env[63531]: DEBUG nova.virt.hardware [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1040.016533] env[63531]: DEBUG nova.virt.hardware [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1040.016681] env[63531]: DEBUG nova.virt.hardware [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1040.016893] env[63531]: DEBUG nova.virt.hardware [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1040.017192] env[63531]: DEBUG nova.virt.hardware [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1040.017394] env[63531]: DEBUG nova.virt.hardware [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1040.017565] env[63531]: DEBUG nova.virt.hardware [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1040.017746] env[63531]: DEBUG nova.virt.hardware [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1040.018687] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7893fc71-b777-4757-9fc9-c31b9562011f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.026375] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5aaa0d2b-3856-4224-9ea6-ecae336bc460 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.144535] env[63531]: DEBUG oslo_concurrency.lockutils [None req-eb1ee106-fb8e-4b9c-876f-c687c3aebb74 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquiring lock "4757625a-8398-4b2d-b6fe-40eea8913068" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.144897] env[63531]: DEBUG oslo_concurrency.lockutils [None req-eb1ee106-fb8e-4b9c-876f-c687c3aebb74 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "4757625a-8398-4b2d-b6fe-40eea8913068" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.307753] env[63531]: DEBUG nova.compute.manager [req-165289c0-a464-4fbc-9f27-fbe038d31c7d req-04ab4408-feb1-4411-9481-d4fcebe43a84 service nova] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] Received event network-changed-b8e8e4e6-84b5-46b4-9b14-65eb203d51cb {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1040.307955] env[63531]: DEBUG nova.compute.manager [req-165289c0-a464-4fbc-9f27-fbe038d31c7d req-04ab4408-feb1-4411-9481-d4fcebe43a84 service nova] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] Refreshing instance network info cache due to event network-changed-b8e8e4e6-84b5-46b4-9b14-65eb203d51cb. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1040.308190] env[63531]: DEBUG oslo_concurrency.lockutils [req-165289c0-a464-4fbc-9f27-fbe038d31c7d req-04ab4408-feb1-4411-9481-d4fcebe43a84 service nova] Acquiring lock "refresh_cache-c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1040.308365] env[63531]: DEBUG oslo_concurrency.lockutils [req-165289c0-a464-4fbc-9f27-fbe038d31c7d req-04ab4408-feb1-4411-9481-d4fcebe43a84 service nova] Acquired lock "refresh_cache-c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1040.308541] env[63531]: DEBUG nova.network.neutron [req-165289c0-a464-4fbc-9f27-fbe038d31c7d req-04ab4408-feb1-4411-9481-d4fcebe43a84 service nova] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] Refreshing network info cache for port b8e8e4e6-84b5-46b4-9b14-65eb203d51cb {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1040.328683] env[63531]: DEBUG nova.network.neutron [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 97d3f108-e299-4ef1-84ca-532cd64905d5] Successfully updated port: b9f690d1-15fa-46cb-a818-c858598dd6f9 {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1040.343779] env[63531]: DEBUG oslo_concurrency.lockutils [None req-df4e3d53-9f91-4cef-a886-ae965b0f89e8 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Lock "ccf00b09-29fe-4a76-a8af-97774acd77de" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.671s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.496133] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118480, 'name': CreateVM_Task, 'duration_secs': 0.310761} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.496532] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1040.497099] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1040.497210] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1040.497488] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1040.497744] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9a2fc1df-5320-42d7-9a08-06c8ed447ef4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.502085] env[63531]: DEBUG oslo_vmware.api [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 1040.502085] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52baeed2-4a74-70f0-c699-cc5ffeef639c" [ 1040.502085] env[63531]: _type = "Task" [ 1040.502085] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.509762] env[63531]: DEBUG oslo_vmware.api [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52baeed2-4a74-70f0-c699-cc5ffeef639c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.523663] env[63531]: DEBUG nova.network.neutron [req-165289c0-a464-4fbc-9f27-fbe038d31c7d req-04ab4408-feb1-4411-9481-d4fcebe43a84 service nova] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] Updated VIF entry in instance network info cache for port b8e8e4e6-84b5-46b4-9b14-65eb203d51cb. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1040.524019] env[63531]: DEBUG nova.network.neutron [req-165289c0-a464-4fbc-9f27-fbe038d31c7d req-04ab4408-feb1-4411-9481-d4fcebe43a84 service nova] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] Updating instance_info_cache with network_info: [{"id": "b8e8e4e6-84b5-46b4-9b14-65eb203d51cb", "address": "fa:16:3e:99:13:e7", "network": {"id": "eac8786d-bc67-4916-8483-b53d9a25efb8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1604989690-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "abdae960d43b47bc95ab41f265d2c85b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db068f71-08cc-42d4-8ab6-17134c1585e5", "external-id": "nsx-vlan-transportzone-721", "segmentation_id": 721, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb8e8e4e6-84", "ovs_interfaceid": "b8e8e4e6-84b5-46b4-9b14-65eb203d51cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1040.541310] env[63531]: DEBUG nova.network.neutron [None req-3afa2054-43b4-4470-afb8-c9c445c20c12 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Updating instance_info_cache with network_info: [{"id": "a926b26b-b99b-49f2-b7dd-6ca158ead545", "address": "fa:16:3e:99:7c:93", "network": {"id": "caf86428-59a5-4e50-9135-2ab8473af64d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1673994696-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.168", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ba8b284854242c392aec5326e996239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa926b26b-b9", "ovs_interfaceid": "a926b26b-b99b-49f2-b7dd-6ca158ead545", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "5f33c6bf-e31f-422e-8611-c4efb3719681", "address": "fa:16:3e:bc:a4:95", "network": {"id": "caf86428-59a5-4e50-9135-2ab8473af64d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1673994696-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ba8b284854242c392aec5326e996239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5f33c6bf-e3", "ovs_interfaceid": "5f33c6bf-e31f-422e-8611-c4efb3719681", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1040.578228] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1040.579188] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb460941-358b-48c1-b8a0-fbcebdd2c2fd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.585160] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c88e5870-9cb0-4b9d-bb8b-b8dd6674abb6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquiring lock "18057005-1ff1-4e3d-bccc-8c89755d6db7" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.585404] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c88e5870-9cb0-4b9d-bb8b-b8dd6674abb6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "18057005-1ff1-4e3d-bccc-8c89755d6db7" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.589067] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ac605ba3-9a33-42b2-a3b2-e72198260c1f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.589289] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ac605ba3-9a33-42b2-a3b2-e72198260c1f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.589474] env[63531]: DEBUG nova.compute.manager [None req-ac605ba3-9a33-42b2-a3b2-e72198260c1f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Going to confirm migration 5 {{(pid=63531) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1040.590631] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1040.591745] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b3620664-1483-47fc-b753-8d20cc3894ac {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.647767] env[63531]: INFO nova.compute.manager [None req-eb1ee106-fb8e-4b9c-876f-c687c3aebb74 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Detaching volume 8537bea9-d5a3-4b76-84eb-8774f4539325 [ 1040.667025] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1040.667025] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1040.667025] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Deleting the datastore file [datastore1] 3fe7a7a8-fc43-4560-b101-227907c91fd0 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1040.667273] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-aa4ba960-6930-4d26-96a4-9fd6a7a96b38 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.676591] env[63531]: DEBUG oslo_vmware.api [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 1040.676591] env[63531]: value = "task-1118482" [ 1040.676591] env[63531]: _type = "Task" [ 1040.676591] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.680592] env[63531]: INFO nova.virt.block_device [None req-eb1ee106-fb8e-4b9c-876f-c687c3aebb74 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Attempting to driver detach volume 8537bea9-d5a3-4b76-84eb-8774f4539325 from mountpoint /dev/sdb [ 1040.680839] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-eb1ee106-fb8e-4b9c-876f-c687c3aebb74 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Volume detach. Driver type: vmdk {{(pid=63531) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1040.681078] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-eb1ee106-fb8e-4b9c-876f-c687c3aebb74 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244788', 'volume_id': '8537bea9-d5a3-4b76-84eb-8774f4539325', 'name': 'volume-8537bea9-d5a3-4b76-84eb-8774f4539325', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4757625a-8398-4b2d-b6fe-40eea8913068', 'attached_at': '', 'detached_at': '', 'volume_id': '8537bea9-d5a3-4b76-84eb-8774f4539325', 'serial': '8537bea9-d5a3-4b76-84eb-8774f4539325'} {{(pid=63531) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1040.681866] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e760e1af-14ae-48a5-9a05-c0449f88c8f6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.690254] env[63531]: DEBUG oslo_vmware.api [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118482, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.706485] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51124dfb-289e-4493-95a0-5047f9f97924 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.713601] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ad2828e-ed78-44ea-a1e9-48608ef85c53 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.734780] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69624606-0052-4ca5-b41d-7910b477bc41 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.749561] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-eb1ee106-fb8e-4b9c-876f-c687c3aebb74 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] The volume has not been displaced from its original location: [datastore2] volume-8537bea9-d5a3-4b76-84eb-8774f4539325/volume-8537bea9-d5a3-4b76-84eb-8774f4539325.vmdk. No consolidation needed. {{(pid=63531) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1040.754656] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-eb1ee106-fb8e-4b9c-876f-c687c3aebb74 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Reconfiguring VM instance instance-0000005a to detach disk 2001 {{(pid=63531) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1040.754931] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f8314839-b170-41e7-b0f3-13737b7d0617 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.773184] env[63531]: DEBUG oslo_vmware.api [None req-eb1ee106-fb8e-4b9c-876f-c687c3aebb74 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for the task: (returnval){ [ 1040.773184] env[63531]: value = "task-1118483" [ 1040.773184] env[63531]: _type = "Task" [ 1040.773184] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.782399] env[63531]: DEBUG oslo_vmware.api [None req-eb1ee106-fb8e-4b9c-876f-c687c3aebb74 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118483, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.834506] env[63531]: DEBUG oslo_concurrency.lockutils [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "refresh_cache-97d3f108-e299-4ef1-84ca-532cd64905d5" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1040.834687] env[63531]: DEBUG oslo_concurrency.lockutils [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquired lock "refresh_cache-97d3f108-e299-4ef1-84ca-532cd64905d5" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1040.834902] env[63531]: DEBUG nova.network.neutron [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 97d3f108-e299-4ef1-84ca-532cd64905d5] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1041.016374] env[63531]: DEBUG oslo_vmware.api [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52baeed2-4a74-70f0-c699-cc5ffeef639c, 'name': SearchDatastore_Task, 'duration_secs': 0.010786} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.016705] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1041.016947] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1041.017205] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1041.017362] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1041.017543] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1041.017812] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-65e88a05-1fd8-463e-b33a-125aaaf6a849 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.026229] env[63531]: DEBUG oslo_concurrency.lockutils [req-165289c0-a464-4fbc-9f27-fbe038d31c7d req-04ab4408-feb1-4411-9481-d4fcebe43a84 service nova] Releasing lock "refresh_cache-c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1041.026635] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1041.026813] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1041.027525] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-019ab8bd-79cb-4930-8103-632256b9f995 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.033780] env[63531]: DEBUG oslo_vmware.api [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 1041.033780] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]524c1ab3-c895-a9b0-8f1b-5871df85d385" [ 1041.033780] env[63531]: _type = "Task" [ 1041.033780] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.041407] env[63531]: DEBUG oslo_vmware.api [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]524c1ab3-c895-a9b0-8f1b-5871df85d385, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.042766] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cec6885c-c8af-4245-84dc-c78230ea7ada {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.045192] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3afa2054-43b4-4470-afb8-c9c445c20c12 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Releasing lock "refresh_cache-e742d3af-5a45-4274-adfb-c6138bf84d6a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1041.045753] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3afa2054-43b4-4470-afb8-c9c445c20c12 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "e742d3af-5a45-4274-adfb-c6138bf84d6a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1041.045912] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3afa2054-43b4-4470-afb8-c9c445c20c12 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquired lock "e742d3af-5a45-4274-adfb-c6138bf84d6a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1041.046621] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ac5cf67-6fff-4d7c-bb27-074358d96667 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.065037] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4965e68-534b-48ae-a683-abe73289ff49 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.068161] env[63531]: DEBUG nova.virt.hardware [None req-3afa2054-43b4-4470-afb8-c9c445c20c12 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1041.068408] env[63531]: DEBUG nova.virt.hardware [None req-3afa2054-43b4-4470-afb8-c9c445c20c12 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1041.068579] env[63531]: DEBUG nova.virt.hardware [None req-3afa2054-43b4-4470-afb8-c9c445c20c12 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1041.068769] env[63531]: DEBUG nova.virt.hardware [None req-3afa2054-43b4-4470-afb8-c9c445c20c12 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1041.068924] env[63531]: DEBUG nova.virt.hardware [None req-3afa2054-43b4-4470-afb8-c9c445c20c12 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1041.069076] env[63531]: DEBUG nova.virt.hardware [None req-3afa2054-43b4-4470-afb8-c9c445c20c12 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1041.069282] env[63531]: DEBUG nova.virt.hardware [None req-3afa2054-43b4-4470-afb8-c9c445c20c12 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1041.069445] env[63531]: DEBUG nova.virt.hardware [None req-3afa2054-43b4-4470-afb8-c9c445c20c12 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1041.069614] env[63531]: DEBUG nova.virt.hardware [None req-3afa2054-43b4-4470-afb8-c9c445c20c12 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1041.069780] env[63531]: DEBUG nova.virt.hardware [None req-3afa2054-43b4-4470-afb8-c9c445c20c12 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1041.069956] env[63531]: DEBUG nova.virt.hardware [None req-3afa2054-43b4-4470-afb8-c9c445c20c12 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1041.075924] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-3afa2054-43b4-4470-afb8-c9c445c20c12 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Reconfiguring VM to attach interface {{(pid=63531) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 1041.076185] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fc82e43d-9757-43f0-8891-4dcae497733b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.094809] env[63531]: DEBUG nova.compute.utils [None req-c88e5870-9cb0-4b9d-bb8b-b8dd6674abb6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1041.095916] env[63531]: DEBUG oslo_vmware.api [None req-3afa2054-43b4-4470-afb8-c9c445c20c12 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for the task: (returnval){ [ 1041.095916] env[63531]: value = "task-1118484" [ 1041.095916] env[63531]: _type = "Task" [ 1041.095916] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.128522] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb6d55b0-84c2-469f-90f9-023f60759738 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.139156] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9259d6c3-b011-41ff-8e1b-81b68e19da49 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.142922] env[63531]: DEBUG oslo_vmware.api [None req-3afa2054-43b4-4470-afb8-c9c445c20c12 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118484, 'name': ReconfigVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.152998] env[63531]: DEBUG nova.compute.provider_tree [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1041.169262] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ac605ba3-9a33-42b2-a3b2-e72198260c1f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "refresh_cache-5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1041.169505] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ac605ba3-9a33-42b2-a3b2-e72198260c1f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquired lock "refresh_cache-5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1041.169719] env[63531]: DEBUG nova.network.neutron [None req-ac605ba3-9a33-42b2-a3b2-e72198260c1f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1041.169955] env[63531]: DEBUG nova.objects.instance [None req-ac605ba3-9a33-42b2-a3b2-e72198260c1f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lazy-loading 'info_cache' on Instance uuid 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1041.187913] env[63531]: DEBUG oslo_vmware.api [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118482, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.133394} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.188213] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1041.188459] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1041.188687] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1041.206932] env[63531]: INFO nova.scheduler.client.report [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Deleted allocations for instance 3fe7a7a8-fc43-4560-b101-227907c91fd0 [ 1041.283482] env[63531]: DEBUG oslo_vmware.api [None req-eb1ee106-fb8e-4b9c-876f-c687c3aebb74 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118483, 'name': ReconfigVM_Task, 'duration_secs': 0.247567} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.283705] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-eb1ee106-fb8e-4b9c-876f-c687c3aebb74 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Reconfigured VM instance instance-0000005a to detach disk 2001 {{(pid=63531) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1041.288236] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-317c3c51-034d-4f18-b095-dc351e304b39 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.303091] env[63531]: DEBUG oslo_vmware.api [None req-eb1ee106-fb8e-4b9c-876f-c687c3aebb74 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for the task: (returnval){ [ 1041.303091] env[63531]: value = "task-1118485" [ 1041.303091] env[63531]: _type = "Task" [ 1041.303091] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.311384] env[63531]: DEBUG oslo_vmware.api [None req-eb1ee106-fb8e-4b9c-876f-c687c3aebb74 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118485, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.368155] env[63531]: DEBUG nova.network.neutron [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 97d3f108-e299-4ef1-84ca-532cd64905d5] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1041.506884] env[63531]: DEBUG nova.network.neutron [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 97d3f108-e299-4ef1-84ca-532cd64905d5] Updating instance_info_cache with network_info: [{"id": "b9f690d1-15fa-46cb-a818-c858598dd6f9", "address": "fa:16:3e:c4:de:f5", "network": {"id": "80641cc0-680b-40ff-8d1b-f5a5f736a88f", "bridge": "br-int", "label": "tempest-ServersTestJSON-666288463-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "63900312a30b4942854f552758c039cf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "26472e27-9835-4f87-ab7f-ca24dfee4e83", "external-id": "nsx-vlan-transportzone-335", "segmentation_id": 335, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb9f690d1-15", "ovs_interfaceid": "b9f690d1-15fa-46cb-a818-c858598dd6f9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1041.544274] env[63531]: DEBUG oslo_vmware.api [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]524c1ab3-c895-a9b0-8f1b-5871df85d385, 'name': SearchDatastore_Task, 'duration_secs': 0.008702} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.545197] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7ce29f40-ac6d-4f34-99e3-178c53f4ed95 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.550513] env[63531]: DEBUG oslo_vmware.api [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 1041.550513] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52ad550e-15a8-2b31-b019-f9b4e325ea21" [ 1041.550513] env[63531]: _type = "Task" [ 1041.550513] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.558081] env[63531]: DEBUG oslo_vmware.api [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52ad550e-15a8-2b31-b019-f9b4e325ea21, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.624581] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c88e5870-9cb0-4b9d-bb8b-b8dd6674abb6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "18057005-1ff1-4e3d-bccc-8c89755d6db7" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.039s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.634544] env[63531]: DEBUG oslo_vmware.api [None req-3afa2054-43b4-4470-afb8-c9c445c20c12 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118484, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.656220] env[63531]: DEBUG nova.scheduler.client.report [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1041.710559] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.813361] env[63531]: DEBUG oslo_vmware.api [None req-eb1ee106-fb8e-4b9c-876f-c687c3aebb74 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118485, 'name': ReconfigVM_Task, 'duration_secs': 0.153712} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.813621] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-eb1ee106-fb8e-4b9c-876f-c687c3aebb74 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244788', 'volume_id': '8537bea9-d5a3-4b76-84eb-8774f4539325', 'name': 'volume-8537bea9-d5a3-4b76-84eb-8774f4539325', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4757625a-8398-4b2d-b6fe-40eea8913068', 'attached_at': '', 'detached_at': '', 'volume_id': '8537bea9-d5a3-4b76-84eb-8774f4539325', 'serial': '8537bea9-d5a3-4b76-84eb-8774f4539325'} {{(pid=63531) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1042.009256] env[63531]: DEBUG oslo_concurrency.lockutils [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Releasing lock "refresh_cache-97d3f108-e299-4ef1-84ca-532cd64905d5" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1042.009645] env[63531]: DEBUG nova.compute.manager [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 97d3f108-e299-4ef1-84ca-532cd64905d5] Instance network_info: |[{"id": "b9f690d1-15fa-46cb-a818-c858598dd6f9", "address": "fa:16:3e:c4:de:f5", "network": {"id": "80641cc0-680b-40ff-8d1b-f5a5f736a88f", "bridge": "br-int", "label": "tempest-ServersTestJSON-666288463-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "63900312a30b4942854f552758c039cf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "26472e27-9835-4f87-ab7f-ca24dfee4e83", "external-id": "nsx-vlan-transportzone-335", "segmentation_id": 335, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb9f690d1-15", "ovs_interfaceid": "b9f690d1-15fa-46cb-a818-c858598dd6f9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1042.010268] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 97d3f108-e299-4ef1-84ca-532cd64905d5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c4:de:f5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '26472e27-9835-4f87-ab7f-ca24dfee4e83', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b9f690d1-15fa-46cb-a818-c858598dd6f9', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1042.018434] env[63531]: DEBUG oslo.service.loopingcall [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1042.018674] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 97d3f108-e299-4ef1-84ca-532cd64905d5] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1042.018906] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b009d453-0cb9-40ad-9392-90c5a684ad61 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.039173] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1042.039173] env[63531]: value = "task-1118486" [ 1042.039173] env[63531]: _type = "Task" [ 1042.039173] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.050041] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118486, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.059345] env[63531]: DEBUG oslo_vmware.api [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52ad550e-15a8-2b31-b019-f9b4e325ea21, 'name': SearchDatastore_Task, 'duration_secs': 0.052323} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.059592] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1042.059854] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9/c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1042.060115] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-09636e24-347e-4b24-b0e7-a9774eea452e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.067072] env[63531]: DEBUG oslo_vmware.api [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 1042.067072] env[63531]: value = "task-1118487" [ 1042.067072] env[63531]: _type = "Task" [ 1042.067072] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.074255] env[63531]: DEBUG oslo_vmware.api [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118487, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.134509] env[63531]: DEBUG oslo_vmware.api [None req-3afa2054-43b4-4470-afb8-c9c445c20c12 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118484, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.160619] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.348s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.161185] env[63531]: DEBUG nova.compute.manager [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] [instance: 3e7b8a01-d930-43cf-9432-0dabc68da20e] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1042.164040] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.453s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.164282] env[63531]: DEBUG nova.objects.instance [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lazy-loading 'resources' on Instance uuid 3fe7a7a8-fc43-4560-b101-227907c91fd0 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1042.359466] env[63531]: DEBUG nova.objects.instance [None req-eb1ee106-fb8e-4b9c-876f-c687c3aebb74 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lazy-loading 'flavor' on Instance uuid 4757625a-8398-4b2d-b6fe-40eea8913068 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1042.507559] env[63531]: DEBUG nova.network.neutron [None req-ac605ba3-9a33-42b2-a3b2-e72198260c1f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Updating instance_info_cache with network_info: [{"id": "91a7c60f-cbcb-4fe3-9bd0-69337882b7ce", "address": "fa:16:3e:46:61:c6", "network": {"id": "bf99b8cb-9efb-4a74-9db1-0f5f0875f26a", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-998205595-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2e99349303e413b9f7896d449cb7dff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc6e6fe1-c4f5-4389-a49f-0978060eebb4", "external-id": "nsx-vlan-transportzone-829", "segmentation_id": 829, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91a7c60f-cb", "ovs_interfaceid": "91a7c60f-cbcb-4fe3-9bd0-69337882b7ce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1042.548899] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118486, 'name': CreateVM_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.576664] env[63531]: DEBUG oslo_vmware.api [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118487, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.635277] env[63531]: DEBUG oslo_vmware.api [None req-3afa2054-43b4-4470-afb8-c9c445c20c12 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118484, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.640276] env[63531]: DEBUG nova.compute.manager [req-a6e36876-517c-409e-9840-fafb808b624e req-ce88ac39-57ba-4269-9376-97a3a2fca52e service nova] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Received event network-changed-5f33c6bf-e31f-422e-8611-c4efb3719681 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1042.640437] env[63531]: DEBUG nova.compute.manager [req-a6e36876-517c-409e-9840-fafb808b624e req-ce88ac39-57ba-4269-9376-97a3a2fca52e service nova] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Refreshing instance network info cache due to event network-changed-5f33c6bf-e31f-422e-8611-c4efb3719681. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1042.640652] env[63531]: DEBUG oslo_concurrency.lockutils [req-a6e36876-517c-409e-9840-fafb808b624e req-ce88ac39-57ba-4269-9376-97a3a2fca52e service nova] Acquiring lock "refresh_cache-e742d3af-5a45-4274-adfb-c6138bf84d6a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1042.640798] env[63531]: DEBUG oslo_concurrency.lockutils [req-a6e36876-517c-409e-9840-fafb808b624e req-ce88ac39-57ba-4269-9376-97a3a2fca52e service nova] Acquired lock "refresh_cache-e742d3af-5a45-4274-adfb-c6138bf84d6a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1042.640961] env[63531]: DEBUG nova.network.neutron [req-a6e36876-517c-409e-9840-fafb808b624e req-ce88ac39-57ba-4269-9376-97a3a2fca52e service nova] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Refreshing network info cache for port 5f33c6bf-e31f-422e-8611-c4efb3719681 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1042.650725] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c88e5870-9cb0-4b9d-bb8b-b8dd6674abb6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquiring lock "18057005-1ff1-4e3d-bccc-8c89755d6db7" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1042.650956] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c88e5870-9cb0-4b9d-bb8b-b8dd6674abb6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "18057005-1ff1-4e3d-bccc-8c89755d6db7" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.651197] env[63531]: INFO nova.compute.manager [None req-c88e5870-9cb0-4b9d-bb8b-b8dd6674abb6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Attaching volume 5c5f7c58-05a8-42fc-9d14-8298686c8120 to /dev/sdb [ 1042.667762] env[63531]: DEBUG nova.compute.utils [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1042.668827] env[63531]: DEBUG nova.objects.instance [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lazy-loading 'numa_topology' on Instance uuid 3fe7a7a8-fc43-4560-b101-227907c91fd0 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1042.669881] env[63531]: DEBUG nova.compute.manager [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] [instance: 3e7b8a01-d930-43cf-9432-0dabc68da20e] Not allocating networking since 'none' was specified. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 1042.683786] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c362bb8-4e02-4c9e-8ea6-6ef9e934e4af {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.690593] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12271456-3be4-418b-8539-4317cb1db186 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.703214] env[63531]: DEBUG nova.virt.block_device [None req-c88e5870-9cb0-4b9d-bb8b-b8dd6674abb6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Updating existing volume attachment record: 8fcd21a2-facc-4548-8cf7-81d8c204fede {{(pid=63531) _volume_attach /opt/stack/nova/nova/virt/block_device.py:679}} [ 1042.867329] env[63531]: DEBUG nova.compute.manager [req-1ecc86e2-fa77-4e5f-a493-03c6b7e5c7bd req-7cfc446d-efb1-457b-8721-0c8ef76c8284 service nova] [instance: 97d3f108-e299-4ef1-84ca-532cd64905d5] Received event network-vif-plugged-b9f690d1-15fa-46cb-a818-c858598dd6f9 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1042.867559] env[63531]: DEBUG oslo_concurrency.lockutils [req-1ecc86e2-fa77-4e5f-a493-03c6b7e5c7bd req-7cfc446d-efb1-457b-8721-0c8ef76c8284 service nova] Acquiring lock "97d3f108-e299-4ef1-84ca-532cd64905d5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1042.867769] env[63531]: DEBUG oslo_concurrency.lockutils [req-1ecc86e2-fa77-4e5f-a493-03c6b7e5c7bd req-7cfc446d-efb1-457b-8721-0c8ef76c8284 service nova] Lock "97d3f108-e299-4ef1-84ca-532cd64905d5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.867941] env[63531]: DEBUG oslo_concurrency.lockutils [req-1ecc86e2-fa77-4e5f-a493-03c6b7e5c7bd req-7cfc446d-efb1-457b-8721-0c8ef76c8284 service nova] Lock "97d3f108-e299-4ef1-84ca-532cd64905d5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.868123] env[63531]: DEBUG nova.compute.manager [req-1ecc86e2-fa77-4e5f-a493-03c6b7e5c7bd req-7cfc446d-efb1-457b-8721-0c8ef76c8284 service nova] [instance: 97d3f108-e299-4ef1-84ca-532cd64905d5] No waiting events found dispatching network-vif-plugged-b9f690d1-15fa-46cb-a818-c858598dd6f9 {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1042.868293] env[63531]: WARNING nova.compute.manager [req-1ecc86e2-fa77-4e5f-a493-03c6b7e5c7bd req-7cfc446d-efb1-457b-8721-0c8ef76c8284 service nova] [instance: 97d3f108-e299-4ef1-84ca-532cd64905d5] Received unexpected event network-vif-plugged-b9f690d1-15fa-46cb-a818-c858598dd6f9 for instance with vm_state building and task_state spawning. [ 1042.868532] env[63531]: DEBUG nova.compute.manager [req-1ecc86e2-fa77-4e5f-a493-03c6b7e5c7bd req-7cfc446d-efb1-457b-8721-0c8ef76c8284 service nova] [instance: 97d3f108-e299-4ef1-84ca-532cd64905d5] Received event network-changed-b9f690d1-15fa-46cb-a818-c858598dd6f9 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1042.868694] env[63531]: DEBUG nova.compute.manager [req-1ecc86e2-fa77-4e5f-a493-03c6b7e5c7bd req-7cfc446d-efb1-457b-8721-0c8ef76c8284 service nova] [instance: 97d3f108-e299-4ef1-84ca-532cd64905d5] Refreshing instance network info cache due to event network-changed-b9f690d1-15fa-46cb-a818-c858598dd6f9. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1042.868879] env[63531]: DEBUG oslo_concurrency.lockutils [req-1ecc86e2-fa77-4e5f-a493-03c6b7e5c7bd req-7cfc446d-efb1-457b-8721-0c8ef76c8284 service nova] Acquiring lock "refresh_cache-97d3f108-e299-4ef1-84ca-532cd64905d5" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1042.869101] env[63531]: DEBUG oslo_concurrency.lockutils [req-1ecc86e2-fa77-4e5f-a493-03c6b7e5c7bd req-7cfc446d-efb1-457b-8721-0c8ef76c8284 service nova] Acquired lock "refresh_cache-97d3f108-e299-4ef1-84ca-532cd64905d5" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1042.869198] env[63531]: DEBUG nova.network.neutron [req-1ecc86e2-fa77-4e5f-a493-03c6b7e5c7bd req-7cfc446d-efb1-457b-8721-0c8ef76c8284 service nova] [instance: 97d3f108-e299-4ef1-84ca-532cd64905d5] Refreshing network info cache for port b9f690d1-15fa-46cb-a818-c858598dd6f9 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1043.010812] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ac605ba3-9a33-42b2-a3b2-e72198260c1f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Releasing lock "refresh_cache-5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1043.011270] env[63531]: DEBUG nova.objects.instance [None req-ac605ba3-9a33-42b2-a3b2-e72198260c1f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lazy-loading 'migration_context' on Instance uuid 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1043.052081] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118486, 'name': CreateVM_Task, 'duration_secs': 0.54096} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.052311] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 97d3f108-e299-4ef1-84ca-532cd64905d5] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1043.053033] env[63531]: DEBUG oslo_concurrency.lockutils [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1043.053273] env[63531]: DEBUG oslo_concurrency.lockutils [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.053651] env[63531]: DEBUG oslo_concurrency.lockutils [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1043.053945] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9dad234e-d33a-4c78-ab89-220522e177ed {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.058607] env[63531]: DEBUG oslo_vmware.api [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 1043.058607] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]525d7093-b5e2-0a7d-a2ba-0b4bc47ecac7" [ 1043.058607] env[63531]: _type = "Task" [ 1043.058607] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.068716] env[63531]: DEBUG oslo_vmware.api [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]525d7093-b5e2-0a7d-a2ba-0b4bc47ecac7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.077691] env[63531]: DEBUG oslo_vmware.api [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118487, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.539669} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.081017] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9/c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1043.081017] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1043.081017] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fee8d24e-2c01-4967-8b27-2505815dd054 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.085166] env[63531]: DEBUG oslo_vmware.api [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 1043.085166] env[63531]: value = "task-1118490" [ 1043.085166] env[63531]: _type = "Task" [ 1043.085166] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.092656] env[63531]: DEBUG oslo_vmware.api [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118490, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.135886] env[63531]: DEBUG oslo_vmware.api [None req-3afa2054-43b4-4470-afb8-c9c445c20c12 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118484, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.171206] env[63531]: DEBUG nova.compute.manager [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] [instance: 3e7b8a01-d930-43cf-9432-0dabc68da20e] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1043.173880] env[63531]: DEBUG nova.objects.base [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Object Instance<3fe7a7a8-fc43-4560-b101-227907c91fd0> lazy-loaded attributes: resources,numa_topology {{(pid=63531) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1043.370871] env[63531]: DEBUG oslo_concurrency.lockutils [None req-eb1ee106-fb8e-4b9c-876f-c687c3aebb74 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "4757625a-8398-4b2d-b6fe-40eea8913068" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.226s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.400360] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb23304d-986c-43fb-aa55-9e87b05ba07d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.407927] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3d8ea3c-437a-4ee4-8dba-1c35cbfcb472 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.438910] env[63531]: DEBUG nova.network.neutron [req-a6e36876-517c-409e-9840-fafb808b624e req-ce88ac39-57ba-4269-9376-97a3a2fca52e service nova] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Updated VIF entry in instance network info cache for port 5f33c6bf-e31f-422e-8611-c4efb3719681. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1043.439379] env[63531]: DEBUG nova.network.neutron [req-a6e36876-517c-409e-9840-fafb808b624e req-ce88ac39-57ba-4269-9376-97a3a2fca52e service nova] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Updating instance_info_cache with network_info: [{"id": "a926b26b-b99b-49f2-b7dd-6ca158ead545", "address": "fa:16:3e:99:7c:93", "network": {"id": "caf86428-59a5-4e50-9135-2ab8473af64d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1673994696-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.168", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ba8b284854242c392aec5326e996239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa926b26b-b9", "ovs_interfaceid": "a926b26b-b99b-49f2-b7dd-6ca158ead545", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "5f33c6bf-e31f-422e-8611-c4efb3719681", "address": "fa:16:3e:bc:a4:95", "network": {"id": "caf86428-59a5-4e50-9135-2ab8473af64d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1673994696-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ba8b284854242c392aec5326e996239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5f33c6bf-e3", "ovs_interfaceid": "5f33c6bf-e31f-422e-8611-c4efb3719681", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1043.443873] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ea2796f-05e1-4635-974a-018f75ae52cc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.449127] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a804328-0827-4991-8774-9a9c09319ef6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.469033] env[63531]: DEBUG nova.compute.provider_tree [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1043.516081] env[63531]: DEBUG nova.objects.base [None req-ac605ba3-9a33-42b2-a3b2-e72198260c1f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Object Instance<5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7> lazy-loaded attributes: info_cache,migration_context {{(pid=63531) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1043.517147] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a65c356-7105-464d-998d-6e6f9435eb57 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.537265] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0dbb42de-3cc4-437f-a680-5f0f3301695a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "3fe7a7a8-fc43-4560-b101-227907c91fd0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.538574] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ce3db076-d04c-470e-9494-9be84679e170 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.543909] env[63531]: DEBUG oslo_vmware.api [None req-ac605ba3-9a33-42b2-a3b2-e72198260c1f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 1043.543909] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5258bc60-52c2-8b73-0614-2a811d91b9f3" [ 1043.543909] env[63531]: _type = "Task" [ 1043.543909] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.552630] env[63531]: DEBUG oslo_vmware.api [None req-ac605ba3-9a33-42b2-a3b2-e72198260c1f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5258bc60-52c2-8b73-0614-2a811d91b9f3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.570995] env[63531]: DEBUG oslo_vmware.api [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]525d7093-b5e2-0a7d-a2ba-0b4bc47ecac7, 'name': SearchDatastore_Task, 'duration_secs': 0.009411} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.571368] env[63531]: DEBUG oslo_concurrency.lockutils [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1043.571709] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 97d3f108-e299-4ef1-84ca-532cd64905d5] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1043.571987] env[63531]: DEBUG oslo_concurrency.lockutils [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1043.572193] env[63531]: DEBUG oslo_concurrency.lockutils [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.572425] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1043.572723] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-83d28fd9-c189-4883-a58c-c8b69cee6d98 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.580859] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1043.581041] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1043.581837] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-34e566e1-3629-43cb-a1a7-2cb2f083f21c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.590269] env[63531]: DEBUG oslo_vmware.api [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 1043.590269] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]529d8ead-048d-e54d-9777-ecf1a61917af" [ 1043.590269] env[63531]: _type = "Task" [ 1043.590269] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.597072] env[63531]: DEBUG oslo_vmware.api [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118490, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.08204} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.597824] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1043.598731] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a8fadb0-01dd-4ab6-be9a-45509c02872d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.604285] env[63531]: DEBUG oslo_vmware.api [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]529d8ead-048d-e54d-9777-ecf1a61917af, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.624607] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] Reconfiguring VM instance instance-00000064 to attach disk [datastore1] c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9/c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1043.624934] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8990edd5-ba0e-47b6-a1e1-5ad89824a0b9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.647907] env[63531]: DEBUG oslo_vmware.api [None req-3afa2054-43b4-4470-afb8-c9c445c20c12 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118484, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.649246] env[63531]: DEBUG oslo_vmware.api [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 1043.649246] env[63531]: value = "task-1118493" [ 1043.649246] env[63531]: _type = "Task" [ 1043.649246] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.659143] env[63531]: DEBUG oslo_vmware.api [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118493, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.928854] env[63531]: DEBUG nova.network.neutron [req-1ecc86e2-fa77-4e5f-a493-03c6b7e5c7bd req-7cfc446d-efb1-457b-8721-0c8ef76c8284 service nova] [instance: 97d3f108-e299-4ef1-84ca-532cd64905d5] Updated VIF entry in instance network info cache for port b9f690d1-15fa-46cb-a818-c858598dd6f9. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1043.929262] env[63531]: DEBUG nova.network.neutron [req-1ecc86e2-fa77-4e5f-a493-03c6b7e5c7bd req-7cfc446d-efb1-457b-8721-0c8ef76c8284 service nova] [instance: 97d3f108-e299-4ef1-84ca-532cd64905d5] Updating instance_info_cache with network_info: [{"id": "b9f690d1-15fa-46cb-a818-c858598dd6f9", "address": "fa:16:3e:c4:de:f5", "network": {"id": "80641cc0-680b-40ff-8d1b-f5a5f736a88f", "bridge": "br-int", "label": "tempest-ServersTestJSON-666288463-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "63900312a30b4942854f552758c039cf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "26472e27-9835-4f87-ab7f-ca24dfee4e83", "external-id": "nsx-vlan-transportzone-335", "segmentation_id": 335, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb9f690d1-15", "ovs_interfaceid": "b9f690d1-15fa-46cb-a818-c858598dd6f9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1043.944798] env[63531]: DEBUG oslo_concurrency.lockutils [req-a6e36876-517c-409e-9840-fafb808b624e req-ce88ac39-57ba-4269-9376-97a3a2fca52e service nova] Releasing lock "refresh_cache-e742d3af-5a45-4274-adfb-c6138bf84d6a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1043.945114] env[63531]: DEBUG nova.compute.manager [req-a6e36876-517c-409e-9840-fafb808b624e req-ce88ac39-57ba-4269-9376-97a3a2fca52e service nova] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Received event network-vif-unplugged-03e99378-6a13-403e-9cb4-856abf141914 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1043.945326] env[63531]: DEBUG oslo_concurrency.lockutils [req-a6e36876-517c-409e-9840-fafb808b624e req-ce88ac39-57ba-4269-9376-97a3a2fca52e service nova] Acquiring lock "3fe7a7a8-fc43-4560-b101-227907c91fd0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.945571] env[63531]: DEBUG oslo_concurrency.lockutils [req-a6e36876-517c-409e-9840-fafb808b624e req-ce88ac39-57ba-4269-9376-97a3a2fca52e service nova] Lock "3fe7a7a8-fc43-4560-b101-227907c91fd0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.945752] env[63531]: DEBUG oslo_concurrency.lockutils [req-a6e36876-517c-409e-9840-fafb808b624e req-ce88ac39-57ba-4269-9376-97a3a2fca52e service nova] Lock "3fe7a7a8-fc43-4560-b101-227907c91fd0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.945928] env[63531]: DEBUG nova.compute.manager [req-a6e36876-517c-409e-9840-fafb808b624e req-ce88ac39-57ba-4269-9376-97a3a2fca52e service nova] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] No waiting events found dispatching network-vif-unplugged-03e99378-6a13-403e-9cb4-856abf141914 {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1043.946126] env[63531]: WARNING nova.compute.manager [req-a6e36876-517c-409e-9840-fafb808b624e req-ce88ac39-57ba-4269-9376-97a3a2fca52e service nova] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Received unexpected event network-vif-unplugged-03e99378-6a13-403e-9cb4-856abf141914 for instance with vm_state shelved_offloaded and task_state None. [ 1043.946298] env[63531]: DEBUG nova.compute.manager [req-a6e36876-517c-409e-9840-fafb808b624e req-ce88ac39-57ba-4269-9376-97a3a2fca52e service nova] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Received event network-changed-03e99378-6a13-403e-9cb4-856abf141914 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1043.946458] env[63531]: DEBUG nova.compute.manager [req-a6e36876-517c-409e-9840-fafb808b624e req-ce88ac39-57ba-4269-9376-97a3a2fca52e service nova] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Refreshing instance network info cache due to event network-changed-03e99378-6a13-403e-9cb4-856abf141914. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1043.946652] env[63531]: DEBUG oslo_concurrency.lockutils [req-a6e36876-517c-409e-9840-fafb808b624e req-ce88ac39-57ba-4269-9376-97a3a2fca52e service nova] Acquiring lock "refresh_cache-3fe7a7a8-fc43-4560-b101-227907c91fd0" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1043.946796] env[63531]: DEBUG oslo_concurrency.lockutils [req-a6e36876-517c-409e-9840-fafb808b624e req-ce88ac39-57ba-4269-9376-97a3a2fca52e service nova] Acquired lock "refresh_cache-3fe7a7a8-fc43-4560-b101-227907c91fd0" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.946959] env[63531]: DEBUG nova.network.neutron [req-a6e36876-517c-409e-9840-fafb808b624e req-ce88ac39-57ba-4269-9376-97a3a2fca52e service nova] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Refreshing network info cache for port 03e99378-6a13-403e-9cb4-856abf141914 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1043.972329] env[63531]: DEBUG nova.scheduler.client.report [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1044.057064] env[63531]: DEBUG oslo_vmware.api [None req-ac605ba3-9a33-42b2-a3b2-e72198260c1f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5258bc60-52c2-8b73-0614-2a811d91b9f3, 'name': SearchDatastore_Task, 'duration_secs': 0.006919} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.057381] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ac605ba3-9a33-42b2-a3b2-e72198260c1f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.101211] env[63531]: DEBUG oslo_vmware.api [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]529d8ead-048d-e54d-9777-ecf1a61917af, 'name': SearchDatastore_Task, 'duration_secs': 0.021766} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.101987] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-868cb8f9-8d3d-461e-8243-ffccf18e70f2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.107498] env[63531]: DEBUG oslo_vmware.api [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 1044.107498] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]529b5178-bb01-5910-4e93-409f202d28d0" [ 1044.107498] env[63531]: _type = "Task" [ 1044.107498] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.114948] env[63531]: DEBUG oslo_vmware.api [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]529b5178-bb01-5910-4e93-409f202d28d0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.147363] env[63531]: DEBUG oslo_vmware.api [None req-3afa2054-43b4-4470-afb8-c9c445c20c12 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118484, 'name': ReconfigVM_Task, 'duration_secs': 2.829817} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.148828] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3afa2054-43b4-4470-afb8-c9c445c20c12 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Releasing lock "e742d3af-5a45-4274-adfb-c6138bf84d6a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.149045] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-3afa2054-43b4-4470-afb8-c9c445c20c12 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Reconfigured VM to attach interface {{(pid=63531) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 1044.162046] env[63531]: DEBUG oslo_vmware.api [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118493, 'name': ReconfigVM_Task, 'duration_secs': 0.284307} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.162325] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] Reconfigured VM instance instance-00000064 to attach disk [datastore1] c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9/c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1044.162918] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1f8e08ac-c529-4d9b-b5f9-8a9f07485a4f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.169887] env[63531]: DEBUG oslo_vmware.api [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 1044.169887] env[63531]: value = "task-1118494" [ 1044.169887] env[63531]: _type = "Task" [ 1044.169887] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.178232] env[63531]: DEBUG oslo_vmware.api [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118494, 'name': Rename_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.182317] env[63531]: DEBUG nova.compute.manager [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] [instance: 3e7b8a01-d930-43cf-9432-0dabc68da20e] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1044.210051] env[63531]: DEBUG nova.virt.hardware [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1044.210375] env[63531]: DEBUG nova.virt.hardware [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1044.210567] env[63531]: DEBUG nova.virt.hardware [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1044.210820] env[63531]: DEBUG nova.virt.hardware [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1044.210989] env[63531]: DEBUG nova.virt.hardware [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1044.211167] env[63531]: DEBUG nova.virt.hardware [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1044.211409] env[63531]: DEBUG nova.virt.hardware [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1044.211571] env[63531]: DEBUG nova.virt.hardware [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1044.211743] env[63531]: DEBUG nova.virt.hardware [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1044.211909] env[63531]: DEBUG nova.virt.hardware [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1044.212113] env[63531]: DEBUG nova.virt.hardware [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1044.213023] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7528d84e-ecfc-48f5-a758-e713285ab13c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.222870] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18b106ad-3e5c-4763-a805-0fe90dae6a4c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.237763] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] [instance: 3e7b8a01-d930-43cf-9432-0dabc68da20e] Instance VIF info [] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1044.243502] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Creating folder: Project (69b85a17a1ae4508928eee1254baeb8a). Parent ref: group-v244585. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1044.243807] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7b014fb2-2712-4896-8e79-30b3aee80bb6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.254318] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Created folder: Project (69b85a17a1ae4508928eee1254baeb8a) in parent group-v244585. [ 1044.254993] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Creating folder: Instances. Parent ref: group-v244793. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1044.255276] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2b1ff7bb-e9e1-42d5-8f6c-573afb8446c5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.264635] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Created folder: Instances in parent group-v244793. [ 1044.264906] env[63531]: DEBUG oslo.service.loopingcall [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1044.265137] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3e7b8a01-d930-43cf-9432-0dabc68da20e] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1044.265372] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-afa96ab5-b9ad-404d-84c2-fc5adfbbab68 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.282058] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1044.282058] env[63531]: value = "task-1118497" [ 1044.282058] env[63531]: _type = "Task" [ 1044.282058] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.290976] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118497, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.432462] env[63531]: DEBUG oslo_concurrency.lockutils [req-1ecc86e2-fa77-4e5f-a493-03c6b7e5c7bd req-7cfc446d-efb1-457b-8721-0c8ef76c8284 service nova] Releasing lock "refresh_cache-97d3f108-e299-4ef1-84ca-532cd64905d5" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.480026] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.316s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.482611] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ac605ba3-9a33-42b2-a3b2-e72198260c1f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.425s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.620152] env[63531]: DEBUG oslo_vmware.api [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]529b5178-bb01-5910-4e93-409f202d28d0, 'name': SearchDatastore_Task, 'duration_secs': 0.010152} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.620613] env[63531]: DEBUG oslo_concurrency.lockutils [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.620891] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 97d3f108-e299-4ef1-84ca-532cd64905d5/97d3f108-e299-4ef1-84ca-532cd64905d5.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1044.621194] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d3955946-01d3-4fea-879c-5601588eb5e4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.628485] env[63531]: DEBUG oslo_vmware.api [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 1044.628485] env[63531]: value = "task-1118498" [ 1044.628485] env[63531]: _type = "Task" [ 1044.628485] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.638161] env[63531]: DEBUG oslo_vmware.api [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118498, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.654749] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3afa2054-43b4-4470-afb8-c9c445c20c12 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "interface-e742d3af-5a45-4274-adfb-c6138bf84d6a-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 8.883s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.658415] env[63531]: DEBUG nova.network.neutron [req-a6e36876-517c-409e-9840-fafb808b624e req-ce88ac39-57ba-4269-9376-97a3a2fca52e service nova] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Updated VIF entry in instance network info cache for port 03e99378-6a13-403e-9cb4-856abf141914. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1044.658415] env[63531]: DEBUG nova.network.neutron [req-a6e36876-517c-409e-9840-fafb808b624e req-ce88ac39-57ba-4269-9376-97a3a2fca52e service nova] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Updating instance_info_cache with network_info: [{"id": "03e99378-6a13-403e-9cb4-856abf141914", "address": "fa:16:3e:3d:57:da", "network": {"id": "ea2d47d3-a502-4afd-9bc5-ee4a37017901", "bridge": null, "label": "tempest-DeleteServersTestJSON-1361282648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94a95667c00947ea9183a6307c569c90", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap03e99378-6a", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1044.658738] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8d83de2b-cc43-41ca-83bd-deb83ce0806c tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquiring lock "4757625a-8398-4b2d-b6fe-40eea8913068" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.658924] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8d83de2b-cc43-41ca-83bd-deb83ce0806c tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "4757625a-8398-4b2d-b6fe-40eea8913068" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.659147] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8d83de2b-cc43-41ca-83bd-deb83ce0806c tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquiring lock "4757625a-8398-4b2d-b6fe-40eea8913068-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.659331] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8d83de2b-cc43-41ca-83bd-deb83ce0806c tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "4757625a-8398-4b2d-b6fe-40eea8913068-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.659514] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8d83de2b-cc43-41ca-83bd-deb83ce0806c tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "4757625a-8398-4b2d-b6fe-40eea8913068-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.661547] env[63531]: INFO nova.compute.manager [None req-8d83de2b-cc43-41ca-83bd-deb83ce0806c tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Terminating instance [ 1044.663372] env[63531]: DEBUG nova.compute.manager [None req-8d83de2b-cc43-41ca-83bd-deb83ce0806c tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1044.663582] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-8d83de2b-cc43-41ca-83bd-deb83ce0806c tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1044.664412] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff17800c-2726-4e67-8251-5ea4ca95a8eb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.675734] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d83de2b-cc43-41ca-83bd-deb83ce0806c tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1044.676294] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b82bb7e0-d1f2-46eb-9aad-8b9f45afa9e6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.681134] env[63531]: DEBUG oslo_vmware.api [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118494, 'name': Rename_Task, 'duration_secs': 0.157278} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.681382] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1044.681597] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-36af998e-8d11-437b-9809-5c490255ba8d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.684183] env[63531]: DEBUG oslo_vmware.api [None req-8d83de2b-cc43-41ca-83bd-deb83ce0806c tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for the task: (returnval){ [ 1044.684183] env[63531]: value = "task-1118499" [ 1044.684183] env[63531]: _type = "Task" [ 1044.684183] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.691485] env[63531]: DEBUG oslo_vmware.api [None req-8d83de2b-cc43-41ca-83bd-deb83ce0806c tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118499, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.693501] env[63531]: DEBUG oslo_vmware.api [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 1044.693501] env[63531]: value = "task-1118500" [ 1044.693501] env[63531]: _type = "Task" [ 1044.693501] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.709710] env[63531]: DEBUG oslo_vmware.api [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118500, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.792208] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118497, 'name': CreateVM_Task, 'duration_secs': 0.313126} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.792377] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3e7b8a01-d930-43cf-9432-0dabc68da20e] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1044.792849] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1044.792986] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.793339] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1044.793603] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4f2c2be2-1939-4e0a-b586-d2874412190b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.798225] env[63531]: DEBUG oslo_vmware.api [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Waiting for the task: (returnval){ [ 1044.798225] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5228e943-8db0-8c4d-69de-39a7e2d7c419" [ 1044.798225] env[63531]: _type = "Task" [ 1044.798225] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.807458] env[63531]: DEBUG oslo_vmware.api [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5228e943-8db0-8c4d-69de-39a7e2d7c419, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.993459] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9b465f1d-4c18-4807-bde7-3f7632026972 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "3fe7a7a8-fc43-4560-b101-227907c91fd0" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 24.625s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.994889] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0dbb42de-3cc4-437f-a680-5f0f3301695a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "3fe7a7a8-fc43-4560-b101-227907c91fd0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 1.458s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.995261] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0dbb42de-3cc4-437f-a680-5f0f3301695a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "3fe7a7a8-fc43-4560-b101-227907c91fd0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.995530] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0dbb42de-3cc4-437f-a680-5f0f3301695a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "3fe7a7a8-fc43-4560-b101-227907c91fd0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.995715] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0dbb42de-3cc4-437f-a680-5f0f3301695a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "3fe7a7a8-fc43-4560-b101-227907c91fd0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.998820] env[63531]: INFO nova.compute.manager [None req-0dbb42de-3cc4-437f-a680-5f0f3301695a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Terminating instance [ 1045.001901] env[63531]: DEBUG nova.compute.manager [None req-0dbb42de-3cc4-437f-a680-5f0f3301695a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1045.002197] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-0dbb42de-3cc4-437f-a680-5f0f3301695a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1045.002577] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5167b3b8-2040-4206-9a56-4d981d3016e7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.013086] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d9e6d5a-eeb4-4ecf-851d-7dc6aaeb7e30 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.050503] env[63531]: WARNING nova.virt.vmwareapi.vmops [None req-0dbb42de-3cc4-437f-a680-5f0f3301695a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3fe7a7a8-fc43-4560-b101-227907c91fd0 could not be found. [ 1045.050735] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-0dbb42de-3cc4-437f-a680-5f0f3301695a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1045.050967] env[63531]: INFO nova.compute.manager [None req-0dbb42de-3cc4-437f-a680-5f0f3301695a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Took 0.05 seconds to destroy the instance on the hypervisor. [ 1045.051256] env[63531]: DEBUG oslo.service.loopingcall [None req-0dbb42de-3cc4-437f-a680-5f0f3301695a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1045.054394] env[63531]: DEBUG nova.compute.manager [-] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1045.054533] env[63531]: DEBUG nova.network.neutron [-] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1045.141859] env[63531]: DEBUG oslo_vmware.api [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118498, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.159362] env[63531]: DEBUG oslo_concurrency.lockutils [req-a6e36876-517c-409e-9840-fafb808b624e req-ce88ac39-57ba-4269-9376-97a3a2fca52e service nova] Releasing lock "refresh_cache-3fe7a7a8-fc43-4560-b101-227907c91fd0" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1045.199299] env[63531]: DEBUG oslo_vmware.api [None req-8d83de2b-cc43-41ca-83bd-deb83ce0806c tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118499, 'name': PowerOffVM_Task, 'duration_secs': 0.225996} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.205410] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d83de2b-cc43-41ca-83bd-deb83ce0806c tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1045.205639] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-8d83de2b-cc43-41ca-83bd-deb83ce0806c tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1045.207414] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c36b1602-1f8f-43fa-b8c4-32449cc984dc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.213795] env[63531]: DEBUG oslo_vmware.api [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118500, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.290727] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-674c6d73-5de9-4529-bc61-f13e25e892bd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.298036] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-8d83de2b-cc43-41ca-83bd-deb83ce0806c tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1045.298334] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-8d83de2b-cc43-41ca-83bd-deb83ce0806c tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Deleting contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1045.298681] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d83de2b-cc43-41ca-83bd-deb83ce0806c tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Deleting the datastore file [datastore2] 4757625a-8398-4b2d-b6fe-40eea8913068 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1045.300230] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cfb34bb2-94b0-46db-94e1-eab527380b45 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.311929] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3080e59-07f6-4ba7-af40-ee06e6f5c735 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.317195] env[63531]: DEBUG oslo_vmware.api [None req-8d83de2b-cc43-41ca-83bd-deb83ce0806c tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for the task: (returnval){ [ 1045.317195] env[63531]: value = "task-1118503" [ 1045.317195] env[63531]: _type = "Task" [ 1045.317195] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.358488] env[63531]: DEBUG oslo_vmware.api [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5228e943-8db0-8c4d-69de-39a7e2d7c419, 'name': SearchDatastore_Task, 'duration_secs': 0.010676} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.359718] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1045.359983] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] [instance: 3e7b8a01-d930-43cf-9432-0dabc68da20e] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1045.360307] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1045.360464] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.360678] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1045.361823] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-561fd1f4-fe4a-4430-83d1-b92ace27a30d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.368029] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-34bba192-f0c6-41b7-8add-ea2b0cb1d468 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.370103] env[63531]: DEBUG oslo_vmware.api [None req-8d83de2b-cc43-41ca-83bd-deb83ce0806c tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118503, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.375877] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b87e7ea-49b8-47c4-ae1d-6c9e29291e9b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.381113] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1045.381361] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1045.382366] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a3985299-fb39-4e16-b3a7-954909c1b9d8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.393014] env[63531]: DEBUG nova.compute.provider_tree [None req-ac605ba3-9a33-42b2-a3b2-e72198260c1f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1045.399361] env[63531]: DEBUG oslo_vmware.api [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Waiting for the task: (returnval){ [ 1045.399361] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]526924b6-278a-705b-cc09-338dfa44943f" [ 1045.399361] env[63531]: _type = "Task" [ 1045.399361] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.406787] env[63531]: DEBUG oslo_vmware.api [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]526924b6-278a-705b-cc09-338dfa44943f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.640317] env[63531]: DEBUG oslo_vmware.api [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118498, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.678899} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.640632] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 97d3f108-e299-4ef1-84ca-532cd64905d5/97d3f108-e299-4ef1-84ca-532cd64905d5.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1045.640819] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 97d3f108-e299-4ef1-84ca-532cd64905d5] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1045.641098] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1183a71b-8a40-4c60-a79a-f43d1649300b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.649846] env[63531]: DEBUG oslo_vmware.api [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 1045.649846] env[63531]: value = "task-1118504" [ 1045.649846] env[63531]: _type = "Task" [ 1045.649846] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.659887] env[63531]: DEBUG oslo_vmware.api [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118504, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.708320] env[63531]: DEBUG oslo_vmware.api [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118500, 'name': PowerOnVM_Task, 'duration_secs': 0.608666} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.708633] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1045.708853] env[63531]: INFO nova.compute.manager [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] Took 8.06 seconds to spawn the instance on the hypervisor. [ 1045.709048] env[63531]: DEBUG nova.compute.manager [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1045.709868] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3477aa28-5040-4ef8-8259-07a5445f7e13 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.827549] env[63531]: DEBUG oslo_vmware.api [None req-8d83de2b-cc43-41ca-83bd-deb83ce0806c tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118503, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.153077} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.827963] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d83de2b-cc43-41ca-83bd-deb83ce0806c tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1045.828199] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-8d83de2b-cc43-41ca-83bd-deb83ce0806c tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Deleted contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1045.828388] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-8d83de2b-cc43-41ca-83bd-deb83ce0806c tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1045.828594] env[63531]: INFO nova.compute.manager [None req-8d83de2b-cc43-41ca-83bd-deb83ce0806c tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Took 1.17 seconds to destroy the instance on the hypervisor. [ 1045.828837] env[63531]: DEBUG oslo.service.loopingcall [None req-8d83de2b-cc43-41ca-83bd-deb83ce0806c tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1045.829044] env[63531]: DEBUG nova.compute.manager [-] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1045.829144] env[63531]: DEBUG nova.network.neutron [-] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1045.898053] env[63531]: DEBUG nova.scheduler.client.report [None req-ac605ba3-9a33-42b2-a3b2-e72198260c1f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1045.924632] env[63531]: DEBUG oslo_vmware.api [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]526924b6-278a-705b-cc09-338dfa44943f, 'name': SearchDatastore_Task, 'duration_secs': 0.011505} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.926913] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-48aaa689-1afd-431d-9083-ab9108c512b8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.933293] env[63531]: DEBUG oslo_vmware.api [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Waiting for the task: (returnval){ [ 1045.933293] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]527eafe7-ac0d-0bc5-075a-d86f49bf2e66" [ 1045.933293] env[63531]: _type = "Task" [ 1045.933293] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.941808] env[63531]: DEBUG oslo_vmware.api [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]527eafe7-ac0d-0bc5-075a-d86f49bf2e66, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.018263] env[63531]: DEBUG nova.network.neutron [-] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1046.162294] env[63531]: DEBUG oslo_vmware.api [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118504, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073559} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.162722] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 97d3f108-e299-4ef1-84ca-532cd64905d5] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1046.163592] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-568606f9-7adc-48dc-b203-bac8c1fbca2c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.188204] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 97d3f108-e299-4ef1-84ca-532cd64905d5] Reconfiguring VM instance instance-00000065 to attach disk [datastore1] 97d3f108-e299-4ef1-84ca-532cd64905d5/97d3f108-e299-4ef1-84ca-532cd64905d5.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1046.189033] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b6b1d3e2-407b-4754-97db-5492a463176b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.210168] env[63531]: DEBUG oslo_vmware.api [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 1046.210168] env[63531]: value = "task-1118505" [ 1046.210168] env[63531]: _type = "Task" [ 1046.210168] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.224208] env[63531]: DEBUG oslo_vmware.api [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118505, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.229057] env[63531]: INFO nova.compute.manager [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] Took 15.12 seconds to build instance. [ 1046.414814] env[63531]: DEBUG oslo_concurrency.lockutils [None req-18cfba05-3c95-4fe8-b4cf-7210aaaf6033 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "interface-e742d3af-5a45-4274-adfb-c6138bf84d6a-4140c945-4976-4856-876d-14507f0a9e17" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.415075] env[63531]: DEBUG oslo_concurrency.lockutils [None req-18cfba05-3c95-4fe8-b4cf-7210aaaf6033 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "interface-e742d3af-5a45-4274-adfb-c6138bf84d6a-4140c945-4976-4856-876d-14507f0a9e17" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.415436] env[63531]: DEBUG nova.objects.instance [None req-18cfba05-3c95-4fe8-b4cf-7210aaaf6033 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lazy-loading 'flavor' on Instance uuid e742d3af-5a45-4274-adfb-c6138bf84d6a {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1046.445901] env[63531]: DEBUG oslo_vmware.api [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]527eafe7-ac0d-0bc5-075a-d86f49bf2e66, 'name': SearchDatastore_Task, 'duration_secs': 0.009366} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.446218] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1046.446497] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] 3e7b8a01-d930-43cf-9432-0dabc68da20e/3e7b8a01-d930-43cf-9432-0dabc68da20e.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1046.446773] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-aef8202d-c709-4b94-8656-410b2aa368a0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.454293] env[63531]: DEBUG oslo_vmware.api [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Waiting for the task: (returnval){ [ 1046.454293] env[63531]: value = "task-1118506" [ 1046.454293] env[63531]: _type = "Task" [ 1046.454293] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.462422] env[63531]: DEBUG oslo_vmware.api [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Task: {'id': task-1118506, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.521100] env[63531]: INFO nova.compute.manager [-] [instance: 3fe7a7a8-fc43-4560-b101-227907c91fd0] Took 1.47 seconds to deallocate network for instance. [ 1046.726326] env[63531]: DEBUG oslo_vmware.api [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118505, 'name': ReconfigVM_Task, 'duration_secs': 0.312089} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.726831] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 97d3f108-e299-4ef1-84ca-532cd64905d5] Reconfigured VM instance instance-00000065 to attach disk [datastore1] 97d3f108-e299-4ef1-84ca-532cd64905d5/97d3f108-e299-4ef1-84ca-532cd64905d5.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1046.727580] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6da52c1f-e0c4-47f5-9f09-a2c6b344ac01 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.729733] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a01fadcb-e702-4ea0-86d8-5f06ab8e568d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.633s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.737469] env[63531]: DEBUG oslo_vmware.api [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 1046.737469] env[63531]: value = "task-1118507" [ 1046.737469] env[63531]: _type = "Task" [ 1046.737469] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.748892] env[63531]: DEBUG oslo_vmware.api [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118507, 'name': Rename_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.920141] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ac605ba3-9a33-42b2-a3b2-e72198260c1f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.437s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.968905] env[63531]: DEBUG oslo_vmware.api [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Task: {'id': task-1118506, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.996144] env[63531]: DEBUG nova.network.neutron [-] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1047.019886] env[63531]: DEBUG oslo_concurrency.lockutils [None req-11bea7a3-32cd-4f38-8cbd-f4972d5dad3e tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Acquiring lock "b7b2b7ac-2321-4cf3-b2ed-4877b893af63" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.020198] env[63531]: DEBUG oslo_concurrency.lockutils [None req-11bea7a3-32cd-4f38-8cbd-f4972d5dad3e tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lock "b7b2b7ac-2321-4cf3-b2ed-4877b893af63" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.056123] env[63531]: DEBUG nova.objects.instance [None req-18cfba05-3c95-4fe8-b4cf-7210aaaf6033 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lazy-loading 'pci_requests' on Instance uuid e742d3af-5a45-4274-adfb-c6138bf84d6a {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1047.247546] env[63531]: DEBUG oslo_vmware.api [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118507, 'name': Rename_Task, 'duration_secs': 0.294615} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.247671] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 97d3f108-e299-4ef1-84ca-532cd64905d5] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1047.248086] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bc31555c-f3bc-4eb0-aae0-35e42a2ccfd9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.252177] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-c88e5870-9cb0-4b9d-bb8b-b8dd6674abb6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Volume attach. Driver type: vmdk {{(pid=63531) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1047.252177] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-c88e5870-9cb0-4b9d-bb8b-b8dd6674abb6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244792', 'volume_id': '5c5f7c58-05a8-42fc-9d14-8298686c8120', 'name': 'volume-5c5f7c58-05a8-42fc-9d14-8298686c8120', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '18057005-1ff1-4e3d-bccc-8c89755d6db7', 'attached_at': '', 'detached_at': '', 'volume_id': '5c5f7c58-05a8-42fc-9d14-8298686c8120', 'serial': '5c5f7c58-05a8-42fc-9d14-8298686c8120'} {{(pid=63531) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1047.252448] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e96ad6d7-c0a7-4af1-b2b9-9654aa63a828 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.256143] env[63531]: DEBUG oslo_vmware.api [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 1047.256143] env[63531]: value = "task-1118508" [ 1047.256143] env[63531]: _type = "Task" [ 1047.256143] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.271629] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2021adb-0a69-475e-8fa2-f5db544ee259 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.277406] env[63531]: DEBUG oslo_vmware.api [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118508, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.300390] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-c88e5870-9cb0-4b9d-bb8b-b8dd6674abb6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Reconfiguring VM instance instance-0000005c to attach disk [datastore2] volume-5c5f7c58-05a8-42fc-9d14-8298686c8120/volume-5c5f7c58-05a8-42fc-9d14-8298686c8120.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1047.300688] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b8ed87a3-993f-4fa6-9c59-e90440207f1e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.318486] env[63531]: DEBUG oslo_vmware.api [None req-c88e5870-9cb0-4b9d-bb8b-b8dd6674abb6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 1047.318486] env[63531]: value = "task-1118509" [ 1047.318486] env[63531]: _type = "Task" [ 1047.318486] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.326275] env[63531]: DEBUG oslo_vmware.api [None req-c88e5870-9cb0-4b9d-bb8b-b8dd6674abb6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118509, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.464620] env[63531]: DEBUG oslo_vmware.api [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Task: {'id': task-1118506, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.528782} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.465227] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] 3e7b8a01-d930-43cf-9432-0dabc68da20e/3e7b8a01-d930-43cf-9432-0dabc68da20e.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1047.465227] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] [instance: 3e7b8a01-d930-43cf-9432-0dabc68da20e] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1047.465349] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-00022ad3-7387-485f-8d56-e6af033b5120 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.471818] env[63531]: DEBUG oslo_vmware.api [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Waiting for the task: (returnval){ [ 1047.471818] env[63531]: value = "task-1118510" [ 1047.471818] env[63531]: _type = "Task" [ 1047.471818] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.480795] env[63531]: DEBUG oslo_vmware.api [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Task: {'id': task-1118510, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.486640] env[63531]: INFO nova.scheduler.client.report [None req-ac605ba3-9a33-42b2-a3b2-e72198260c1f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Deleted allocation for migration 2e83b091-f95e-45ca-9edc-3ab1f0e9a94a [ 1047.499534] env[63531]: INFO nova.compute.manager [-] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Took 1.67 seconds to deallocate network for instance. [ 1047.524104] env[63531]: DEBUG nova.compute.utils [None req-11bea7a3-32cd-4f38-8cbd-f4972d5dad3e tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1047.548648] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0dbb42de-3cc4-437f-a680-5f0f3301695a tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "3fe7a7a8-fc43-4560-b101-227907c91fd0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.554s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.558277] env[63531]: DEBUG nova.objects.base [None req-18cfba05-3c95-4fe8-b4cf-7210aaaf6033 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=63531) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1047.558643] env[63531]: DEBUG nova.network.neutron [None req-18cfba05-3c95-4fe8-b4cf-7210aaaf6033 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1047.624900] env[63531]: DEBUG nova.policy [None req-18cfba05-3c95-4fe8-b4cf-7210aaaf6033 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '36cb2f0de54f417384fa1fb2210de4e4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8ba8b284854242c392aec5326e996239', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 1047.766906] env[63531]: DEBUG oslo_vmware.api [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118508, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.832032] env[63531]: DEBUG oslo_vmware.api [None req-c88e5870-9cb0-4b9d-bb8b-b8dd6674abb6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118509, 'name': ReconfigVM_Task, 'duration_secs': 0.389983} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.832032] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-c88e5870-9cb0-4b9d-bb8b-b8dd6674abb6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Reconfigured VM instance instance-0000005c to attach disk [datastore2] volume-5c5f7c58-05a8-42fc-9d14-8298686c8120/volume-5c5f7c58-05a8-42fc-9d14-8298686c8120.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1047.839478] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-725f9d86-e08b-4248-94d5-cf8ffbce0d12 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.862857] env[63531]: DEBUG oslo_vmware.api [None req-c88e5870-9cb0-4b9d-bb8b-b8dd6674abb6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 1047.862857] env[63531]: value = "task-1118511" [ 1047.862857] env[63531]: _type = "Task" [ 1047.862857] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.871743] env[63531]: DEBUG oslo_vmware.api [None req-c88e5870-9cb0-4b9d-bb8b-b8dd6674abb6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118511, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.981473] env[63531]: DEBUG oslo_vmware.api [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Task: {'id': task-1118510, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.09412} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.981473] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] [instance: 3e7b8a01-d930-43cf-9432-0dabc68da20e] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1047.982357] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-452aacec-0018-4d06-a667-5b44a5cea837 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.996185] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ac605ba3-9a33-42b2-a3b2-e72198260c1f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 7.407s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.005594] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] [instance: 3e7b8a01-d930-43cf-9432-0dabc68da20e] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] 3e7b8a01-d930-43cf-9432-0dabc68da20e/3e7b8a01-d930-43cf-9432-0dabc68da20e.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1048.006797] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8d83de2b-cc43-41ca-83bd-deb83ce0806c tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1048.007137] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8d83de2b-cc43-41ca-83bd-deb83ce0806c tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.007478] env[63531]: DEBUG nova.objects.instance [None req-8d83de2b-cc43-41ca-83bd-deb83ce0806c tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lazy-loading 'resources' on Instance uuid 4757625a-8398-4b2d-b6fe-40eea8913068 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1048.008414] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f1c076b0-9c95-46f0-9269-53a20ea67543 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.027366] env[63531]: DEBUG oslo_concurrency.lockutils [None req-11bea7a3-32cd-4f38-8cbd-f4972d5dad3e tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lock "b7b2b7ac-2321-4cf3-b2ed-4877b893af63" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.034415] env[63531]: DEBUG oslo_vmware.api [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Waiting for the task: (returnval){ [ 1048.034415] env[63531]: value = "task-1118512" [ 1048.034415] env[63531]: _type = "Task" [ 1048.034415] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.042570] env[63531]: DEBUG oslo_vmware.api [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Task: {'id': task-1118512, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.210463] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27d0a0b9-f99c-47ab-b941-bc8758e46561 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.217847] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ecc05ff-a382-4b78-9deb-00ffe788c18e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.247811] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec5e4e72-5d15-423c-9b5b-c5df101545a8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.255856] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c6845f7-bac0-461d-b28a-73f2260a829e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.272273] env[63531]: DEBUG nova.compute.provider_tree [None req-8d83de2b-cc43-41ca-83bd-deb83ce0806c tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1048.276599] env[63531]: DEBUG oslo_vmware.api [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118508, 'name': PowerOnVM_Task, 'duration_secs': 0.611622} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.276599] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 97d3f108-e299-4ef1-84ca-532cd64905d5] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1048.276785] env[63531]: INFO nova.compute.manager [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 97d3f108-e299-4ef1-84ca-532cd64905d5] Took 8.29 seconds to spawn the instance on the hypervisor. [ 1048.276966] env[63531]: DEBUG nova.compute.manager [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 97d3f108-e299-4ef1-84ca-532cd64905d5] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1048.277721] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b48b386-6d9c-46a9-8d92-5100c1c7dea6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.374546] env[63531]: DEBUG oslo_vmware.api [None req-c88e5870-9cb0-4b9d-bb8b-b8dd6674abb6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118511, 'name': ReconfigVM_Task, 'duration_secs': 0.305415} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.374866] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-c88e5870-9cb0-4b9d-bb8b-b8dd6674abb6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244792', 'volume_id': '5c5f7c58-05a8-42fc-9d14-8298686c8120', 'name': 'volume-5c5f7c58-05a8-42fc-9d14-8298686c8120', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '18057005-1ff1-4e3d-bccc-8c89755d6db7', 'attached_at': '', 'detached_at': '', 'volume_id': '5c5f7c58-05a8-42fc-9d14-8298686c8120', 'serial': '5c5f7c58-05a8-42fc-9d14-8298686c8120'} {{(pid=63531) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1048.544585] env[63531]: DEBUG oslo_vmware.api [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Task: {'id': task-1118512, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.777498] env[63531]: DEBUG nova.scheduler.client.report [None req-8d83de2b-cc43-41ca-83bd-deb83ce0806c tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1048.793031] env[63531]: INFO nova.compute.manager [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 97d3f108-e299-4ef1-84ca-532cd64905d5] Took 16.97 seconds to build instance. [ 1049.030497] env[63531]: DEBUG nova.compute.manager [req-3ba1c2b7-0ec4-490a-967a-a3eb4fa46542 req-aa820c12-542a-406f-924c-55b307af0bf2 service nova] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Received event network-vif-deleted-b37a092f-e03c-417f-bac1-3b2ebb5253d1 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1049.030770] env[63531]: DEBUG nova.compute.manager [req-3ba1c2b7-0ec4-490a-967a-a3eb4fa46542 req-aa820c12-542a-406f-924c-55b307af0bf2 service nova] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] Received event network-changed-b8e8e4e6-84b5-46b4-9b14-65eb203d51cb {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1049.031016] env[63531]: DEBUG nova.compute.manager [req-3ba1c2b7-0ec4-490a-967a-a3eb4fa46542 req-aa820c12-542a-406f-924c-55b307af0bf2 service nova] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] Refreshing instance network info cache due to event network-changed-b8e8e4e6-84b5-46b4-9b14-65eb203d51cb. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1049.031410] env[63531]: DEBUG oslo_concurrency.lockutils [req-3ba1c2b7-0ec4-490a-967a-a3eb4fa46542 req-aa820c12-542a-406f-924c-55b307af0bf2 service nova] Acquiring lock "refresh_cache-c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1049.031617] env[63531]: DEBUG oslo_concurrency.lockutils [req-3ba1c2b7-0ec4-490a-967a-a3eb4fa46542 req-aa820c12-542a-406f-924c-55b307af0bf2 service nova] Acquired lock "refresh_cache-c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1049.031885] env[63531]: DEBUG nova.network.neutron [req-3ba1c2b7-0ec4-490a-967a-a3eb4fa46542 req-aa820c12-542a-406f-924c-55b307af0bf2 service nova] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] Refreshing network info cache for port b8e8e4e6-84b5-46b4-9b14-65eb203d51cb {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1049.049234] env[63531]: DEBUG oslo_vmware.api [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Task: {'id': task-1118512, 'name': ReconfigVM_Task, 'duration_secs': 0.603764} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.049500] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] [instance: 3e7b8a01-d930-43cf-9432-0dabc68da20e] Reconfigured VM instance instance-00000066 to attach disk [datastore2] 3e7b8a01-d930-43cf-9432-0dabc68da20e/3e7b8a01-d930-43cf-9432-0dabc68da20e.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1049.050141] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-40b0bec9-fa9f-4112-9934-be9f363815c5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.057605] env[63531]: DEBUG oslo_vmware.api [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Waiting for the task: (returnval){ [ 1049.057605] env[63531]: value = "task-1118513" [ 1049.057605] env[63531]: _type = "Task" [ 1049.057605] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.066669] env[63531]: DEBUG oslo_vmware.api [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Task: {'id': task-1118513, 'name': Rename_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.084263] env[63531]: DEBUG oslo_concurrency.lockutils [None req-11bea7a3-32cd-4f38-8cbd-f4972d5dad3e tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Acquiring lock "b7b2b7ac-2321-4cf3-b2ed-4877b893af63" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.084511] env[63531]: DEBUG oslo_concurrency.lockutils [None req-11bea7a3-32cd-4f38-8cbd-f4972d5dad3e tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lock "b7b2b7ac-2321-4cf3-b2ed-4877b893af63" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1049.084753] env[63531]: INFO nova.compute.manager [None req-11bea7a3-32cd-4f38-8cbd-f4972d5dad3e tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Attaching volume a34e5b41-b9fc-4c5a-967e-b3e601dde931 to /dev/sdb [ 1049.121904] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-790491ae-9040-45eb-97dc-fd734a9b5db8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.133587] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39b17e60-5bc3-4bdd-aad8-b97984af44e4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.147934] env[63531]: DEBUG nova.virt.block_device [None req-11bea7a3-32cd-4f38-8cbd-f4972d5dad3e tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Updating existing volume attachment record: 5012926f-72c9-4d08-868c-1972ba4e8eaa {{(pid=63531) _volume_attach /opt/stack/nova/nova/virt/block_device.py:679}} [ 1049.167173] env[63531]: DEBUG nova.network.neutron [None req-18cfba05-3c95-4fe8-b4cf-7210aaaf6033 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Successfully updated port: 4140c945-4976-4856-876d-14507f0a9e17 {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1049.282455] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8d83de2b-cc43-41ca-83bd-deb83ce0806c tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.275s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.296453] env[63531]: DEBUG oslo_concurrency.lockutils [None req-949e02cd-0a7b-47a9-a7d6-5317b4570f77 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "97d3f108-e299-4ef1-84ca-532cd64905d5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.487s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.305492] env[63531]: INFO nova.scheduler.client.report [None req-8d83de2b-cc43-41ca-83bd-deb83ce0806c tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Deleted allocations for instance 4757625a-8398-4b2d-b6fe-40eea8913068 [ 1049.410409] env[63531]: DEBUG nova.objects.instance [None req-c88e5870-9cb0-4b9d-bb8b-b8dd6674abb6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lazy-loading 'flavor' on Instance uuid 18057005-1ff1-4e3d-bccc-8c89755d6db7 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1049.567100] env[63531]: DEBUG oslo_vmware.api [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Task: {'id': task-1118513, 'name': Rename_Task, 'duration_secs': 0.23115} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.567384] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] [instance: 3e7b8a01-d930-43cf-9432-0dabc68da20e] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1049.567636] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c129a894-6887-4eff-969a-673b7c59276a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.573859] env[63531]: DEBUG oslo_vmware.api [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Waiting for the task: (returnval){ [ 1049.573859] env[63531]: value = "task-1118515" [ 1049.573859] env[63531]: _type = "Task" [ 1049.573859] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.584807] env[63531]: DEBUG oslo_vmware.api [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Task: {'id': task-1118515, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.604413] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0e8b28d-c0dc-41b1-b98b-a77c5c3f4e22 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquiring lock "879924d4-b465-4102-a0e3-c7b2be7ef08b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.604669] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0e8b28d-c0dc-41b1-b98b-a77c5c3f4e22 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Lock "879924d4-b465-4102-a0e3-c7b2be7ef08b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1049.604892] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0e8b28d-c0dc-41b1-b98b-a77c5c3f4e22 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquiring lock "879924d4-b465-4102-a0e3-c7b2be7ef08b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.605108] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0e8b28d-c0dc-41b1-b98b-a77c5c3f4e22 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Lock "879924d4-b465-4102-a0e3-c7b2be7ef08b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1049.605257] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0e8b28d-c0dc-41b1-b98b-a77c5c3f4e22 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Lock "879924d4-b465-4102-a0e3-c7b2be7ef08b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.607525] env[63531]: INFO nova.compute.manager [None req-c0e8b28d-c0dc-41b1-b98b-a77c5c3f4e22 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Terminating instance [ 1049.609414] env[63531]: DEBUG nova.compute.manager [None req-c0e8b28d-c0dc-41b1-b98b-a77c5c3f4e22 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1049.609667] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c0e8b28d-c0dc-41b1-b98b-a77c5c3f4e22 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1049.610533] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-287c3b61-f222-4520-bd0e-9512808482cd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.618908] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0e8b28d-c0dc-41b1-b98b-a77c5c3f4e22 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1049.618908] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-51d9ec25-0be2-4f1e-984b-0388591f7cae {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.625884] env[63531]: DEBUG oslo_vmware.api [None req-c0e8b28d-c0dc-41b1-b98b-a77c5c3f4e22 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 1049.625884] env[63531]: value = "task-1118516" [ 1049.625884] env[63531]: _type = "Task" [ 1049.625884] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.634680] env[63531]: DEBUG oslo_vmware.api [None req-c0e8b28d-c0dc-41b1-b98b-a77c5c3f4e22 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118516, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.669274] env[63531]: DEBUG oslo_concurrency.lockutils [None req-18cfba05-3c95-4fe8-b4cf-7210aaaf6033 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "refresh_cache-e742d3af-5a45-4274-adfb-c6138bf84d6a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1049.669453] env[63531]: DEBUG oslo_concurrency.lockutils [None req-18cfba05-3c95-4fe8-b4cf-7210aaaf6033 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquired lock "refresh_cache-e742d3af-5a45-4274-adfb-c6138bf84d6a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1049.669635] env[63531]: DEBUG nova.network.neutron [None req-18cfba05-3c95-4fe8-b4cf-7210aaaf6033 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1049.817971] env[63531]: DEBUG oslo_concurrency.lockutils [None req-8d83de2b-cc43-41ca-83bd-deb83ce0806c tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "4757625a-8398-4b2d-b6fe-40eea8913068" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.159s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.864905] env[63531]: DEBUG nova.network.neutron [req-3ba1c2b7-0ec4-490a-967a-a3eb4fa46542 req-aa820c12-542a-406f-924c-55b307af0bf2 service nova] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] Updated VIF entry in instance network info cache for port b8e8e4e6-84b5-46b4-9b14-65eb203d51cb. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1049.865319] env[63531]: DEBUG nova.network.neutron [req-3ba1c2b7-0ec4-490a-967a-a3eb4fa46542 req-aa820c12-542a-406f-924c-55b307af0bf2 service nova] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] Updating instance_info_cache with network_info: [{"id": "b8e8e4e6-84b5-46b4-9b14-65eb203d51cb", "address": "fa:16:3e:99:13:e7", "network": {"id": "eac8786d-bc67-4916-8483-b53d9a25efb8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1604989690-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "abdae960d43b47bc95ab41f265d2c85b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db068f71-08cc-42d4-8ab6-17134c1585e5", "external-id": "nsx-vlan-transportzone-721", "segmentation_id": 721, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb8e8e4e6-84", "ovs_interfaceid": "b8e8e4e6-84b5-46b4-9b14-65eb203d51cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1049.918176] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c88e5870-9cb0-4b9d-bb8b-b8dd6674abb6 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "18057005-1ff1-4e3d-bccc-8c89755d6db7" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.267s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.085600] env[63531]: DEBUG oslo_vmware.api [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Task: {'id': task-1118515, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.136485] env[63531]: DEBUG oslo_vmware.api [None req-c0e8b28d-c0dc-41b1-b98b-a77c5c3f4e22 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118516, 'name': PowerOffVM_Task, 'duration_secs': 0.217082} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.136768] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0e8b28d-c0dc-41b1-b98b-a77c5c3f4e22 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1050.136950] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c0e8b28d-c0dc-41b1-b98b-a77c5c3f4e22 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1050.137215] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3c2f233c-e373-444f-9747-fae190002177 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.217736] env[63531]: WARNING nova.network.neutron [None req-18cfba05-3c95-4fe8-b4cf-7210aaaf6033 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] caf86428-59a5-4e50-9135-2ab8473af64d already exists in list: networks containing: ['caf86428-59a5-4e50-9135-2ab8473af64d']. ignoring it [ 1050.217890] env[63531]: WARNING nova.network.neutron [None req-18cfba05-3c95-4fe8-b4cf-7210aaaf6033 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] caf86428-59a5-4e50-9135-2ab8473af64d already exists in list: networks containing: ['caf86428-59a5-4e50-9135-2ab8473af64d']. ignoring it [ 1050.221414] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c0e8b28d-c0dc-41b1-b98b-a77c5c3f4e22 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1050.221665] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c0e8b28d-c0dc-41b1-b98b-a77c5c3f4e22 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Deleting contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1050.221937] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0e8b28d-c0dc-41b1-b98b-a77c5c3f4e22 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Deleting the datastore file [datastore2] 879924d4-b465-4102-a0e3-c7b2be7ef08b {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1050.222192] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-64b8d57e-59d9-4c99-8f4e-1310bea88daf {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.227774] env[63531]: DEBUG oslo_vmware.api [None req-c0e8b28d-c0dc-41b1-b98b-a77c5c3f4e22 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 1050.227774] env[63531]: value = "task-1118518" [ 1050.227774] env[63531]: _type = "Task" [ 1050.227774] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.235452] env[63531]: DEBUG oslo_vmware.api [None req-c0e8b28d-c0dc-41b1-b98b-a77c5c3f4e22 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118518, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.367973] env[63531]: DEBUG oslo_concurrency.lockutils [req-3ba1c2b7-0ec4-490a-967a-a3eb4fa46542 req-aa820c12-542a-406f-924c-55b307af0bf2 service nova] Releasing lock "refresh_cache-c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1050.585812] env[63531]: DEBUG oslo_vmware.api [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Task: {'id': task-1118515, 'name': PowerOnVM_Task, 'duration_secs': 0.544462} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.587475] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] [instance: 3e7b8a01-d930-43cf-9432-0dabc68da20e] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1050.587475] env[63531]: INFO nova.compute.manager [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] [instance: 3e7b8a01-d930-43cf-9432-0dabc68da20e] Took 6.40 seconds to spawn the instance on the hypervisor. [ 1050.587475] env[63531]: DEBUG nova.compute.manager [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] [instance: 3e7b8a01-d930-43cf-9432-0dabc68da20e] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1050.587475] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0d801f4-d83b-4246-8865-2ce553a361e6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.637394] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "8f903d83-adbd-4f71-a6f1-037807746248" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1050.637638] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "8f903d83-adbd-4f71-a6f1-037807746248" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1050.740087] env[63531]: DEBUG oslo_vmware.api [None req-c0e8b28d-c0dc-41b1-b98b-a77c5c3f4e22 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118518, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.265025} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.740368] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0e8b28d-c0dc-41b1-b98b-a77c5c3f4e22 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1050.740556] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c0e8b28d-c0dc-41b1-b98b-a77c5c3f4e22 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Deleted contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1050.740738] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c0e8b28d-c0dc-41b1-b98b-a77c5c3f4e22 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1050.740936] env[63531]: INFO nova.compute.manager [None req-c0e8b28d-c0dc-41b1-b98b-a77c5c3f4e22 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1050.741200] env[63531]: DEBUG oslo.service.loopingcall [None req-c0e8b28d-c0dc-41b1-b98b-a77c5c3f4e22 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1050.741395] env[63531]: DEBUG nova.compute.manager [-] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1050.741493] env[63531]: DEBUG nova.network.neutron [-] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1050.863588] env[63531]: DEBUG nova.network.neutron [None req-18cfba05-3c95-4fe8-b4cf-7210aaaf6033 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Updating instance_info_cache with network_info: [{"id": "a926b26b-b99b-49f2-b7dd-6ca158ead545", "address": "fa:16:3e:99:7c:93", "network": {"id": "caf86428-59a5-4e50-9135-2ab8473af64d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1673994696-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.168", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ba8b284854242c392aec5326e996239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa926b26b-b9", "ovs_interfaceid": "a926b26b-b99b-49f2-b7dd-6ca158ead545", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "5f33c6bf-e31f-422e-8611-c4efb3719681", "address": "fa:16:3e:bc:a4:95", "network": {"id": "caf86428-59a5-4e50-9135-2ab8473af64d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1673994696-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ba8b284854242c392aec5326e996239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5f33c6bf-e3", "ovs_interfaceid": "5f33c6bf-e31f-422e-8611-c4efb3719681", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "4140c945-4976-4856-876d-14507f0a9e17", "address": "fa:16:3e:34:05:a3", "network": {"id": "caf86428-59a5-4e50-9135-2ab8473af64d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1673994696-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ba8b284854242c392aec5326e996239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4140c945-49", "ovs_interfaceid": "4140c945-4976-4856-876d-14507f0a9e17", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1051.108350] env[63531]: INFO nova.compute.manager [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] [instance: 3e7b8a01-d930-43cf-9432-0dabc68da20e] Took 12.60 seconds to build instance. [ 1051.142029] env[63531]: DEBUG nova.compute.manager [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 8f903d83-adbd-4f71-a6f1-037807746248] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1051.367055] env[63531]: DEBUG oslo_concurrency.lockutils [None req-18cfba05-3c95-4fe8-b4cf-7210aaaf6033 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Releasing lock "refresh_cache-e742d3af-5a45-4274-adfb-c6138bf84d6a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1051.367990] env[63531]: DEBUG oslo_concurrency.lockutils [None req-18cfba05-3c95-4fe8-b4cf-7210aaaf6033 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "e742d3af-5a45-4274-adfb-c6138bf84d6a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1051.368244] env[63531]: DEBUG oslo_concurrency.lockutils [None req-18cfba05-3c95-4fe8-b4cf-7210aaaf6033 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquired lock "e742d3af-5a45-4274-adfb-c6138bf84d6a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1051.369522] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d2352c0-9027-4533-95d8-2ff8185d61cb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.388050] env[63531]: DEBUG nova.virt.hardware [None req-18cfba05-3c95-4fe8-b4cf-7210aaaf6033 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1051.388357] env[63531]: DEBUG nova.virt.hardware [None req-18cfba05-3c95-4fe8-b4cf-7210aaaf6033 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1051.388543] env[63531]: DEBUG nova.virt.hardware [None req-18cfba05-3c95-4fe8-b4cf-7210aaaf6033 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1051.388845] env[63531]: DEBUG nova.virt.hardware [None req-18cfba05-3c95-4fe8-b4cf-7210aaaf6033 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1051.388943] env[63531]: DEBUG nova.virt.hardware [None req-18cfba05-3c95-4fe8-b4cf-7210aaaf6033 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1051.389574] env[63531]: DEBUG nova.virt.hardware [None req-18cfba05-3c95-4fe8-b4cf-7210aaaf6033 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1051.389574] env[63531]: DEBUG nova.virt.hardware [None req-18cfba05-3c95-4fe8-b4cf-7210aaaf6033 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1051.389574] env[63531]: DEBUG nova.virt.hardware [None req-18cfba05-3c95-4fe8-b4cf-7210aaaf6033 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1051.389746] env[63531]: DEBUG nova.virt.hardware [None req-18cfba05-3c95-4fe8-b4cf-7210aaaf6033 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1051.389891] env[63531]: DEBUG nova.virt.hardware [None req-18cfba05-3c95-4fe8-b4cf-7210aaaf6033 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1051.390813] env[63531]: DEBUG nova.virt.hardware [None req-18cfba05-3c95-4fe8-b4cf-7210aaaf6033 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1051.396668] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-18cfba05-3c95-4fe8-b4cf-7210aaaf6033 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Reconfiguring VM to attach interface {{(pid=63531) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 1051.397066] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e2a430ef-f51b-4deb-97b3-dbc76ac405e4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.420286] env[63531]: DEBUG oslo_vmware.api [None req-18cfba05-3c95-4fe8-b4cf-7210aaaf6033 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for the task: (returnval){ [ 1051.420286] env[63531]: value = "task-1118520" [ 1051.420286] env[63531]: _type = "Task" [ 1051.420286] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.429227] env[63531]: DEBUG oslo_vmware.api [None req-18cfba05-3c95-4fe8-b4cf-7210aaaf6033 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118520, 'name': ReconfigVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.431726] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b92663e6-8900-4f7f-bc86-533aa67ecc8f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.431993] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b92663e6-8900-4f7f-bc86-533aa67ecc8f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.432167] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b92663e6-8900-4f7f-bc86-533aa67ecc8f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.432351] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b92663e6-8900-4f7f-bc86-533aa67ecc8f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.432518] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b92663e6-8900-4f7f-bc86-533aa67ecc8f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1051.434615] env[63531]: INFO nova.compute.manager [None req-b92663e6-8900-4f7f-bc86-533aa67ecc8f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Terminating instance [ 1051.436306] env[63531]: DEBUG nova.compute.manager [None req-b92663e6-8900-4f7f-bc86-533aa67ecc8f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1051.436500] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-b92663e6-8900-4f7f-bc86-533aa67ecc8f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1051.437277] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7da641fe-db2e-46c2-b75f-19aab9ca4d78 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.444767] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-b92663e6-8900-4f7f-bc86-533aa67ecc8f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1051.447973] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b8808767-dbd7-4b83-8817-17b9a94f63ef {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.450362] env[63531]: DEBUG oslo_vmware.api [None req-b92663e6-8900-4f7f-bc86-533aa67ecc8f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 1051.450362] env[63531]: value = "task-1118521" [ 1051.450362] env[63531]: _type = "Task" [ 1051.450362] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.459621] env[63531]: DEBUG oslo_vmware.api [None req-b92663e6-8900-4f7f-bc86-533aa67ecc8f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118521, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.561533] env[63531]: DEBUG nova.network.neutron [-] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1051.612138] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0e2eb460-f432-410e-9b72-1ce02e5cb693 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Lock "3e7b8a01-d930-43cf-9432-0dabc68da20e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.110s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1051.615262] env[63531]: DEBUG nova.compute.manager [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Stashing vm_state: active {{(pid=63531) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1051.666408] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.666949] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.668341] env[63531]: INFO nova.compute.claims [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 8f903d83-adbd-4f71-a6f1-037807746248] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1051.686877] env[63531]: DEBUG oslo_concurrency.lockutils [None req-aa491dbe-51b5-432e-ab71-8cdd1877a2f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "97d3f108-e299-4ef1-84ca-532cd64905d5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.687108] env[63531]: DEBUG oslo_concurrency.lockutils [None req-aa491dbe-51b5-432e-ab71-8cdd1877a2f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "97d3f108-e299-4ef1-84ca-532cd64905d5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.687319] env[63531]: DEBUG oslo_concurrency.lockutils [None req-aa491dbe-51b5-432e-ab71-8cdd1877a2f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "97d3f108-e299-4ef1-84ca-532cd64905d5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.687505] env[63531]: DEBUG oslo_concurrency.lockutils [None req-aa491dbe-51b5-432e-ab71-8cdd1877a2f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "97d3f108-e299-4ef1-84ca-532cd64905d5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.687675] env[63531]: DEBUG oslo_concurrency.lockutils [None req-aa491dbe-51b5-432e-ab71-8cdd1877a2f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "97d3f108-e299-4ef1-84ca-532cd64905d5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1051.690386] env[63531]: INFO nova.compute.manager [None req-aa491dbe-51b5-432e-ab71-8cdd1877a2f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 97d3f108-e299-4ef1-84ca-532cd64905d5] Terminating instance [ 1051.692222] env[63531]: DEBUG nova.compute.manager [None req-aa491dbe-51b5-432e-ab71-8cdd1877a2f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 97d3f108-e299-4ef1-84ca-532cd64905d5] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1051.692414] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-aa491dbe-51b5-432e-ab71-8cdd1877a2f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 97d3f108-e299-4ef1-84ca-532cd64905d5] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1051.693253] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edae81dd-a79d-4b86-905a-869effacdcb1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.699084] env[63531]: DEBUG nova.compute.manager [req-1dbec15b-f094-47dc-8b9a-9ae4b565d213 req-e8c7030b-4e97-4bd1-b682-372416ea03f3 service nova] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Received event network-vif-plugged-4140c945-4976-4856-876d-14507f0a9e17 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1051.699084] env[63531]: DEBUG oslo_concurrency.lockutils [req-1dbec15b-f094-47dc-8b9a-9ae4b565d213 req-e8c7030b-4e97-4bd1-b682-372416ea03f3 service nova] Acquiring lock "e742d3af-5a45-4274-adfb-c6138bf84d6a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.699084] env[63531]: DEBUG oslo_concurrency.lockutils [req-1dbec15b-f094-47dc-8b9a-9ae4b565d213 req-e8c7030b-4e97-4bd1-b682-372416ea03f3 service nova] Lock "e742d3af-5a45-4274-adfb-c6138bf84d6a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.699084] env[63531]: DEBUG oslo_concurrency.lockutils [req-1dbec15b-f094-47dc-8b9a-9ae4b565d213 req-e8c7030b-4e97-4bd1-b682-372416ea03f3 service nova] Lock "e742d3af-5a45-4274-adfb-c6138bf84d6a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1051.699084] env[63531]: DEBUG nova.compute.manager [req-1dbec15b-f094-47dc-8b9a-9ae4b565d213 req-e8c7030b-4e97-4bd1-b682-372416ea03f3 service nova] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] No waiting events found dispatching network-vif-plugged-4140c945-4976-4856-876d-14507f0a9e17 {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1051.699084] env[63531]: WARNING nova.compute.manager [req-1dbec15b-f094-47dc-8b9a-9ae4b565d213 req-e8c7030b-4e97-4bd1-b682-372416ea03f3 service nova] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Received unexpected event network-vif-plugged-4140c945-4976-4856-876d-14507f0a9e17 for instance with vm_state active and task_state None. [ 1051.699728] env[63531]: DEBUG nova.compute.manager [req-1dbec15b-f094-47dc-8b9a-9ae4b565d213 req-e8c7030b-4e97-4bd1-b682-372416ea03f3 service nova] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Received event network-changed-4140c945-4976-4856-876d-14507f0a9e17 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1051.699728] env[63531]: DEBUG nova.compute.manager [req-1dbec15b-f094-47dc-8b9a-9ae4b565d213 req-e8c7030b-4e97-4bd1-b682-372416ea03f3 service nova] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Refreshing instance network info cache due to event network-changed-4140c945-4976-4856-876d-14507f0a9e17. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1051.699728] env[63531]: DEBUG oslo_concurrency.lockutils [req-1dbec15b-f094-47dc-8b9a-9ae4b565d213 req-e8c7030b-4e97-4bd1-b682-372416ea03f3 service nova] Acquiring lock "refresh_cache-e742d3af-5a45-4274-adfb-c6138bf84d6a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1051.699728] env[63531]: DEBUG oslo_concurrency.lockutils [req-1dbec15b-f094-47dc-8b9a-9ae4b565d213 req-e8c7030b-4e97-4bd1-b682-372416ea03f3 service nova] Acquired lock "refresh_cache-e742d3af-5a45-4274-adfb-c6138bf84d6a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1051.699728] env[63531]: DEBUG nova.network.neutron [req-1dbec15b-f094-47dc-8b9a-9ae4b565d213 req-e8c7030b-4e97-4bd1-b682-372416ea03f3 service nova] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Refreshing network info cache for port 4140c945-4976-4856-876d-14507f0a9e17 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1051.709315] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa491dbe-51b5-432e-ab71-8cdd1877a2f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 97d3f108-e299-4ef1-84ca-532cd64905d5] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1051.709315] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5e609403-a7be-452f-960a-b91bd302b36f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.713994] env[63531]: DEBUG oslo_vmware.api [None req-aa491dbe-51b5-432e-ab71-8cdd1877a2f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 1051.713994] env[63531]: value = "task-1118522" [ 1051.713994] env[63531]: _type = "Task" [ 1051.713994] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.723525] env[63531]: DEBUG oslo_vmware.api [None req-aa491dbe-51b5-432e-ab71-8cdd1877a2f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118522, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.932116] env[63531]: DEBUG oslo_vmware.api [None req-18cfba05-3c95-4fe8-b4cf-7210aaaf6033 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118520, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.960566] env[63531]: DEBUG oslo_vmware.api [None req-b92663e6-8900-4f7f-bc86-533aa67ecc8f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118521, 'name': PowerOffVM_Task, 'duration_secs': 0.193215} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.962380] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-b92663e6-8900-4f7f-bc86-533aa67ecc8f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1051.962380] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-b92663e6-8900-4f7f-bc86-533aa67ecc8f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1051.962380] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fb8ffe00-a849-4329-bb3f-4f83e29aa5a1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.064638] env[63531]: INFO nova.compute.manager [-] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Took 1.32 seconds to deallocate network for instance. [ 1052.129031] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-b92663e6-8900-4f7f-bc86-533aa67ecc8f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1052.129279] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-b92663e6-8900-4f7f-bc86-533aa67ecc8f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Deleting contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1052.129409] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-b92663e6-8900-4f7f-bc86-533aa67ecc8f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Deleting the datastore file [datastore2] 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1052.129789] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cfd3b7ec-64dd-41d4-a23e-8d04eef2cb06 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.136479] env[63531]: DEBUG oslo_vmware.api [None req-b92663e6-8900-4f7f-bc86-533aa67ecc8f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 1052.136479] env[63531]: value = "task-1118525" [ 1052.136479] env[63531]: _type = "Task" [ 1052.136479] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.137388] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.146017] env[63531]: DEBUG oslo_vmware.api [None req-b92663e6-8900-4f7f-bc86-533aa67ecc8f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118525, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.226282] env[63531]: DEBUG oslo_vmware.api [None req-aa491dbe-51b5-432e-ab71-8cdd1877a2f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118522, 'name': PowerOffVM_Task, 'duration_secs': 0.200428} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.226675] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa491dbe-51b5-432e-ab71-8cdd1877a2f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 97d3f108-e299-4ef1-84ca-532cd64905d5] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1052.227581] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-aa491dbe-51b5-432e-ab71-8cdd1877a2f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 97d3f108-e299-4ef1-84ca-532cd64905d5] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1052.227996] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-188ef51a-89a0-4ba5-940a-048954d40f5c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.301736] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "85058146-ba4c-4e9a-bbe7-048ed56167ab" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.302065] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "85058146-ba4c-4e9a-bbe7-048ed56167ab" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.307986] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-aa491dbe-51b5-432e-ab71-8cdd1877a2f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 97d3f108-e299-4ef1-84ca-532cd64905d5] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1052.308339] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-aa491dbe-51b5-432e-ab71-8cdd1877a2f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 97d3f108-e299-4ef1-84ca-532cd64905d5] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1052.308415] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa491dbe-51b5-432e-ab71-8cdd1877a2f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Deleting the datastore file [datastore1] 97d3f108-e299-4ef1-84ca-532cd64905d5 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1052.308694] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5658d07a-5525-48e7-a3f1-0d928b5ade77 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.315764] env[63531]: DEBUG oslo_vmware.api [None req-aa491dbe-51b5-432e-ab71-8cdd1877a2f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 1052.315764] env[63531]: value = "task-1118527" [ 1052.315764] env[63531]: _type = "Task" [ 1052.315764] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.330825] env[63531]: DEBUG oslo_vmware.api [None req-aa491dbe-51b5-432e-ab71-8cdd1877a2f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118527, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.431594] env[63531]: DEBUG oslo_vmware.api [None req-18cfba05-3c95-4fe8-b4cf-7210aaaf6033 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118520, 'name': ReconfigVM_Task, 'duration_secs': 0.716835} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.431594] env[63531]: DEBUG oslo_concurrency.lockutils [None req-18cfba05-3c95-4fe8-b4cf-7210aaaf6033 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Releasing lock "e742d3af-5a45-4274-adfb-c6138bf84d6a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1052.431594] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-18cfba05-3c95-4fe8-b4cf-7210aaaf6033 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Reconfigured VM to attach interface {{(pid=63531) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 1052.454379] env[63531]: DEBUG nova.network.neutron [req-1dbec15b-f094-47dc-8b9a-9ae4b565d213 req-e8c7030b-4e97-4bd1-b682-372416ea03f3 service nova] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Updated VIF entry in instance network info cache for port 4140c945-4976-4856-876d-14507f0a9e17. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1052.454927] env[63531]: DEBUG nova.network.neutron [req-1dbec15b-f094-47dc-8b9a-9ae4b565d213 req-e8c7030b-4e97-4bd1-b682-372416ea03f3 service nova] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Updating instance_info_cache with network_info: [{"id": "a926b26b-b99b-49f2-b7dd-6ca158ead545", "address": "fa:16:3e:99:7c:93", "network": {"id": "caf86428-59a5-4e50-9135-2ab8473af64d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1673994696-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.168", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ba8b284854242c392aec5326e996239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa926b26b-b9", "ovs_interfaceid": "a926b26b-b99b-49f2-b7dd-6ca158ead545", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "5f33c6bf-e31f-422e-8611-c4efb3719681", "address": "fa:16:3e:bc:a4:95", "network": {"id": "caf86428-59a5-4e50-9135-2ab8473af64d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1673994696-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ba8b284854242c392aec5326e996239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5f33c6bf-e3", "ovs_interfaceid": "5f33c6bf-e31f-422e-8611-c4efb3719681", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "4140c945-4976-4856-876d-14507f0a9e17", "address": "fa:16:3e:34:05:a3", "network": {"id": "caf86428-59a5-4e50-9135-2ab8473af64d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1673994696-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ba8b284854242c392aec5326e996239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4140c945-49", "ovs_interfaceid": "4140c945-4976-4856-876d-14507f0a9e17", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1052.571882] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0e8b28d-c0dc-41b1-b98b-a77c5c3f4e22 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.649978] env[63531]: DEBUG oslo_vmware.api [None req-b92663e6-8900-4f7f-bc86-533aa67ecc8f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118525, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.18462} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.650294] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-b92663e6-8900-4f7f-bc86-533aa67ecc8f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1052.650500] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-b92663e6-8900-4f7f-bc86-533aa67ecc8f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Deleted contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1052.650699] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-b92663e6-8900-4f7f-bc86-533aa67ecc8f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1052.651017] env[63531]: INFO nova.compute.manager [None req-b92663e6-8900-4f7f-bc86-533aa67ecc8f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Took 1.21 seconds to destroy the instance on the hypervisor. [ 1052.651229] env[63531]: DEBUG oslo.service.loopingcall [None req-b92663e6-8900-4f7f-bc86-533aa67ecc8f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1052.651445] env[63531]: DEBUG nova.compute.manager [-] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1052.651542] env[63531]: DEBUG nova.network.neutron [-] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1052.808209] env[63531]: DEBUG nova.compute.manager [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 85058146-ba4c-4e9a-bbe7-048ed56167ab] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1052.829021] env[63531]: DEBUG oslo_vmware.api [None req-aa491dbe-51b5-432e-ab71-8cdd1877a2f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118527, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.283544} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.831982] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa491dbe-51b5-432e-ab71-8cdd1877a2f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1052.832280] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-aa491dbe-51b5-432e-ab71-8cdd1877a2f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 97d3f108-e299-4ef1-84ca-532cd64905d5] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1052.832544] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-aa491dbe-51b5-432e-ab71-8cdd1877a2f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 97d3f108-e299-4ef1-84ca-532cd64905d5] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1052.835284] env[63531]: INFO nova.compute.manager [None req-aa491dbe-51b5-432e-ab71-8cdd1877a2f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 97d3f108-e299-4ef1-84ca-532cd64905d5] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1052.835284] env[63531]: DEBUG oslo.service.loopingcall [None req-aa491dbe-51b5-432e-ab71-8cdd1877a2f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1052.835284] env[63531]: DEBUG nova.compute.manager [-] [instance: 97d3f108-e299-4ef1-84ca-532cd64905d5] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1052.835284] env[63531]: DEBUG nova.network.neutron [-] [instance: 97d3f108-e299-4ef1-84ca-532cd64905d5] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1052.920602] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f68ce3e6-cd7c-4e9b-8bb8-6d1c44d9b5ae {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.928785] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcec9c98-a82d-4c55-b770-82f90da200ff {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.936235] env[63531]: DEBUG oslo_concurrency.lockutils [None req-18cfba05-3c95-4fe8-b4cf-7210aaaf6033 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "interface-e742d3af-5a45-4274-adfb-c6138bf84d6a-4140c945-4976-4856-876d-14507f0a9e17" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.521s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.973700] env[63531]: DEBUG oslo_concurrency.lockutils [req-1dbec15b-f094-47dc-8b9a-9ae4b565d213 req-e8c7030b-4e97-4bd1-b682-372416ea03f3 service nova] Releasing lock "refresh_cache-e742d3af-5a45-4274-adfb-c6138bf84d6a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1052.974500] env[63531]: DEBUG nova.compute.manager [None req-4a218e88-e7de-443e-9bd8-2856e300ffba tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] [instance: 3e7b8a01-d930-43cf-9432-0dabc68da20e] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1052.977209] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3634d1b6-a479-449b-bdd0-c20f4b0868a2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.982981] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f254b28a-4787-4951-9b90-4da2ff933011 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.996683] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-077213c3-89c0-4df0-a1ca-0f3d4143e0f8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.018748] env[63531]: DEBUG nova.compute.provider_tree [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1053.186619] env[63531]: DEBUG oslo_concurrency.lockutils [None req-584426ce-5cc1-4420-87bd-4a33f7084720 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Acquiring lock "3e7b8a01-d930-43cf-9432-0dabc68da20e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1053.186870] env[63531]: DEBUG oslo_concurrency.lockutils [None req-584426ce-5cc1-4420-87bd-4a33f7084720 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Lock "3e7b8a01-d930-43cf-9432-0dabc68da20e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1053.187116] env[63531]: DEBUG oslo_concurrency.lockutils [None req-584426ce-5cc1-4420-87bd-4a33f7084720 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Acquiring lock "3e7b8a01-d930-43cf-9432-0dabc68da20e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1053.187373] env[63531]: DEBUG oslo_concurrency.lockutils [None req-584426ce-5cc1-4420-87bd-4a33f7084720 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Lock "3e7b8a01-d930-43cf-9432-0dabc68da20e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1053.187558] env[63531]: DEBUG oslo_concurrency.lockutils [None req-584426ce-5cc1-4420-87bd-4a33f7084720 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Lock "3e7b8a01-d930-43cf-9432-0dabc68da20e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.189789] env[63531]: INFO nova.compute.manager [None req-584426ce-5cc1-4420-87bd-4a33f7084720 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] [instance: 3e7b8a01-d930-43cf-9432-0dabc68da20e] Terminating instance [ 1053.191730] env[63531]: DEBUG oslo_concurrency.lockutils [None req-584426ce-5cc1-4420-87bd-4a33f7084720 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Acquiring lock "refresh_cache-3e7b8a01-d930-43cf-9432-0dabc68da20e" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1053.191898] env[63531]: DEBUG oslo_concurrency.lockutils [None req-584426ce-5cc1-4420-87bd-4a33f7084720 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Acquired lock "refresh_cache-3e7b8a01-d930-43cf-9432-0dabc68da20e" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1053.192162] env[63531]: DEBUG nova.network.neutron [None req-584426ce-5cc1-4420-87bd-4a33f7084720 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] [instance: 3e7b8a01-d930-43cf-9432-0dabc68da20e] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1053.337504] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1053.504243] env[63531]: INFO nova.compute.manager [None req-4a218e88-e7de-443e-9bd8-2856e300ffba tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] [instance: 3e7b8a01-d930-43cf-9432-0dabc68da20e] instance snapshotting [ 1053.504856] env[63531]: DEBUG nova.objects.instance [None req-4a218e88-e7de-443e-9bd8-2856e300ffba tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Lazy-loading 'flavor' on Instance uuid 3e7b8a01-d930-43cf-9432-0dabc68da20e {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1053.521287] env[63531]: DEBUG nova.scheduler.client.report [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1053.540818] env[63531]: DEBUG nova.network.neutron [-] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1053.697997] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-11bea7a3-32cd-4f38-8cbd-f4972d5dad3e tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Volume attach. Driver type: vmdk {{(pid=63531) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1053.698277] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-11bea7a3-32cd-4f38-8cbd-f4972d5dad3e tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244796', 'volume_id': 'a34e5b41-b9fc-4c5a-967e-b3e601dde931', 'name': 'volume-a34e5b41-b9fc-4c5a-967e-b3e601dde931', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b7b2b7ac-2321-4cf3-b2ed-4877b893af63', 'attached_at': '', 'detached_at': '', 'volume_id': 'a34e5b41-b9fc-4c5a-967e-b3e601dde931', 'serial': 'a34e5b41-b9fc-4c5a-967e-b3e601dde931'} {{(pid=63531) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1053.699243] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67a4cec0-1c6c-4be0-81c4-02554f6cc9bc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.717491] env[63531]: DEBUG nova.network.neutron [None req-584426ce-5cc1-4420-87bd-4a33f7084720 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] [instance: 3e7b8a01-d930-43cf-9432-0dabc68da20e] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1053.718941] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea185ff0-821b-4044-bbb1-058cd50b15eb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.738644] env[63531]: DEBUG nova.network.neutron [-] [instance: 97d3f108-e299-4ef1-84ca-532cd64905d5] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1053.746918] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-11bea7a3-32cd-4f38-8cbd-f4972d5dad3e tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Reconfiguring VM instance instance-0000005d to attach disk [datastore2] volume-a34e5b41-b9fc-4c5a-967e-b3e601dde931/volume-a34e5b41-b9fc-4c5a-967e-b3e601dde931.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1053.751111] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7f8d203b-238d-4312-ae5b-19a65310bcf9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.765445] env[63531]: DEBUG nova.compute.manager [req-ec7eb5d5-66cc-45c4-8280-19ce2f866da1 req-af96a89a-c6a4-4a19-ad26-b94c8ea7c784 service nova] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Received event network-vif-deleted-d5d578b0-1f4a-4a46-8fac-1e625d48c2bc {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1053.765665] env[63531]: DEBUG nova.compute.manager [req-ec7eb5d5-66cc-45c4-8280-19ce2f866da1 req-af96a89a-c6a4-4a19-ad26-b94c8ea7c784 service nova] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Received event network-vif-deleted-91a7c60f-cbcb-4fe3-9bd0-69337882b7ce {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1053.765935] env[63531]: DEBUG nova.compute.manager [req-ec7eb5d5-66cc-45c4-8280-19ce2f866da1 req-af96a89a-c6a4-4a19-ad26-b94c8ea7c784 service nova] [instance: 97d3f108-e299-4ef1-84ca-532cd64905d5] Received event network-vif-deleted-b9f690d1-15fa-46cb-a818-c858598dd6f9 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1053.765999] env[63531]: INFO nova.compute.manager [req-ec7eb5d5-66cc-45c4-8280-19ce2f866da1 req-af96a89a-c6a4-4a19-ad26-b94c8ea7c784 service nova] [instance: 97d3f108-e299-4ef1-84ca-532cd64905d5] Neutron deleted interface b9f690d1-15fa-46cb-a818-c858598dd6f9; detaching it from the instance and deleting it from the info cache [ 1053.766187] env[63531]: DEBUG nova.network.neutron [req-ec7eb5d5-66cc-45c4-8280-19ce2f866da1 req-af96a89a-c6a4-4a19-ad26-b94c8ea7c784 service nova] [instance: 97d3f108-e299-4ef1-84ca-532cd64905d5] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1053.776368] env[63531]: DEBUG oslo_vmware.api [None req-11bea7a3-32cd-4f38-8cbd-f4972d5dad3e tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Waiting for the task: (returnval){ [ 1053.776368] env[63531]: value = "task-1118528" [ 1053.776368] env[63531]: _type = "Task" [ 1053.776368] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.784418] env[63531]: DEBUG oslo_vmware.api [None req-11bea7a3-32cd-4f38-8cbd-f4972d5dad3e tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118528, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.817199] env[63531]: DEBUG nova.network.neutron [None req-584426ce-5cc1-4420-87bd-4a33f7084720 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] [instance: 3e7b8a01-d930-43cf-9432-0dabc68da20e] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1053.899022] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquiring lock "89f19669-53a9-4425-a3c6-d5f8873b707e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1053.899271] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "89f19669-53a9-4425-a3c6-d5f8873b707e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1054.011609] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcfa7d09-580f-4539-a5e4-848a63d36c54 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.028886] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.362s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1054.029474] env[63531]: DEBUG nova.compute.manager [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 8f903d83-adbd-4f71-a6f1-037807746248] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1054.032582] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 1.895s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1054.034176] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b66d0b1-8a5e-4ffe-8a71-d53fb6c81206 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.043432] env[63531]: INFO nova.compute.manager [-] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Took 1.39 seconds to deallocate network for instance. [ 1054.248880] env[63531]: INFO nova.compute.manager [-] [instance: 97d3f108-e299-4ef1-84ca-532cd64905d5] Took 1.41 seconds to deallocate network for instance. [ 1054.269571] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fe0e0502-4738-4f56-926c-8d5e4505f34c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.283319] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cbd4883-0daa-44c9-8882-b984e07e148e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.298057] env[63531]: DEBUG oslo_concurrency.lockutils [None req-09413a86-09cb-400a-aa04-8dbb3cff6da3 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "interface-e742d3af-5a45-4274-adfb-c6138bf84d6a-5f33c6bf-e31f-422e-8611-c4efb3719681" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1054.298057] env[63531]: DEBUG oslo_concurrency.lockutils [None req-09413a86-09cb-400a-aa04-8dbb3cff6da3 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "interface-e742d3af-5a45-4274-adfb-c6138bf84d6a-5f33c6bf-e31f-422e-8611-c4efb3719681" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1054.302461] env[63531]: DEBUG oslo_vmware.api [None req-11bea7a3-32cd-4f38-8cbd-f4972d5dad3e tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118528, 'name': ReconfigVM_Task, 'duration_secs': 0.399846} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.302930] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-11bea7a3-32cd-4f38-8cbd-f4972d5dad3e tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Reconfigured VM instance instance-0000005d to attach disk [datastore2] volume-a34e5b41-b9fc-4c5a-967e-b3e601dde931/volume-a34e5b41-b9fc-4c5a-967e-b3e601dde931.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1054.317111] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7f2e5478-ccb1-449e-8094-9aa61f3511b1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.327241] env[63531]: DEBUG nova.compute.manager [req-ec7eb5d5-66cc-45c4-8280-19ce2f866da1 req-af96a89a-c6a4-4a19-ad26-b94c8ea7c784 service nova] [instance: 97d3f108-e299-4ef1-84ca-532cd64905d5] Detach interface failed, port_id=b9f690d1-15fa-46cb-a818-c858598dd6f9, reason: Instance 97d3f108-e299-4ef1-84ca-532cd64905d5 could not be found. {{(pid=63531) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1054.327907] env[63531]: DEBUG oslo_concurrency.lockutils [None req-584426ce-5cc1-4420-87bd-4a33f7084720 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Releasing lock "refresh_cache-3e7b8a01-d930-43cf-9432-0dabc68da20e" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1054.328329] env[63531]: DEBUG nova.compute.manager [None req-584426ce-5cc1-4420-87bd-4a33f7084720 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] [instance: 3e7b8a01-d930-43cf-9432-0dabc68da20e] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1054.328520] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-584426ce-5cc1-4420-87bd-4a33f7084720 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] [instance: 3e7b8a01-d930-43cf-9432-0dabc68da20e] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1054.329338] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb0e5b33-1e75-4fce-9fc0-da96385efe50 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.340203] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-584426ce-5cc1-4420-87bd-4a33f7084720 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] [instance: 3e7b8a01-d930-43cf-9432-0dabc68da20e] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1054.342132] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6610f6bc-6593-4e19-950c-0dfce2cdab81 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.344660] env[63531]: DEBUG oslo_vmware.api [None req-11bea7a3-32cd-4f38-8cbd-f4972d5dad3e tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Waiting for the task: (returnval){ [ 1054.344660] env[63531]: value = "task-1118529" [ 1054.344660] env[63531]: _type = "Task" [ 1054.344660] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.352330] env[63531]: DEBUG oslo_vmware.api [None req-584426ce-5cc1-4420-87bd-4a33f7084720 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Waiting for the task: (returnval){ [ 1054.352330] env[63531]: value = "task-1118530" [ 1054.352330] env[63531]: _type = "Task" [ 1054.352330] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.362775] env[63531]: DEBUG oslo_vmware.api [None req-11bea7a3-32cd-4f38-8cbd-f4972d5dad3e tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118529, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.378535] env[63531]: DEBUG oslo_vmware.api [None req-584426ce-5cc1-4420-87bd-4a33f7084720 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Task: {'id': task-1118530, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.401567] env[63531]: DEBUG nova.compute.manager [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1054.534589] env[63531]: DEBUG nova.compute.utils [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1054.536736] env[63531]: DEBUG nova.compute.manager [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 8f903d83-adbd-4f71-a6f1-037807746248] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1054.536909] env[63531]: DEBUG nova.network.neutron [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 8f903d83-adbd-4f71-a6f1-037807746248] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1054.542015] env[63531]: INFO nova.compute.claims [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1054.545691] env[63531]: DEBUG nova.compute.manager [None req-4a218e88-e7de-443e-9bd8-2856e300ffba tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] [instance: 3e7b8a01-d930-43cf-9432-0dabc68da20e] Instance disappeared during snapshot {{(pid=63531) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 1054.548806] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b92663e6-8900-4f7f-bc86-533aa67ecc8f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1054.723234] env[63531]: DEBUG nova.compute.manager [None req-4a218e88-e7de-443e-9bd8-2856e300ffba tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] [instance: 3e7b8a01-d930-43cf-9432-0dabc68da20e] Found 0 images (rotation: 2) {{(pid=63531) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 1054.734977] env[63531]: DEBUG nova.policy [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0372c22d1e9e4f66a0417643dab4a2ec', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '94a95667c00947ea9183a6307c569c90', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 1054.756840] env[63531]: DEBUG oslo_concurrency.lockutils [None req-aa491dbe-51b5-432e-ab71-8cdd1877a2f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1054.803894] env[63531]: DEBUG oslo_concurrency.lockutils [None req-09413a86-09cb-400a-aa04-8dbb3cff6da3 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "e742d3af-5a45-4274-adfb-c6138bf84d6a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1054.803894] env[63531]: DEBUG oslo_concurrency.lockutils [None req-09413a86-09cb-400a-aa04-8dbb3cff6da3 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquired lock "e742d3af-5a45-4274-adfb-c6138bf84d6a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1054.806673] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b14a8050-8998-4b8c-a0d0-a15d0878deed {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.834058] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d43aa7fc-29cd-4cda-8f49-1b5ca8eeaafd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.874445] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-09413a86-09cb-400a-aa04-8dbb3cff6da3 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Reconfiguring VM to detach interface {{(pid=63531) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 1054.880749] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8a66c473-e548-428a-b552-04a141e739cf {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.901450] env[63531]: DEBUG oslo_vmware.api [None req-584426ce-5cc1-4420-87bd-4a33f7084720 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Task: {'id': task-1118530, 'name': PowerOffVM_Task, 'duration_secs': 0.425164} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.907460] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-584426ce-5cc1-4420-87bd-4a33f7084720 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] [instance: 3e7b8a01-d930-43cf-9432-0dabc68da20e] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1054.907748] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-584426ce-5cc1-4420-87bd-4a33f7084720 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] [instance: 3e7b8a01-d930-43cf-9432-0dabc68da20e] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1054.908441] env[63531]: DEBUG oslo_vmware.api [None req-09413a86-09cb-400a-aa04-8dbb3cff6da3 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for the task: (returnval){ [ 1054.908441] env[63531]: value = "task-1118531" [ 1054.908441] env[63531]: _type = "Task" [ 1054.908441] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.911601] env[63531]: DEBUG oslo_vmware.api [None req-11bea7a3-32cd-4f38-8cbd-f4972d5dad3e tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118529, 'name': ReconfigVM_Task, 'duration_secs': 0.180936} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.911999] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0d42005e-641c-422e-a5fd-341943ebea16 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.915234] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-11bea7a3-32cd-4f38-8cbd-f4972d5dad3e tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244796', 'volume_id': 'a34e5b41-b9fc-4c5a-967e-b3e601dde931', 'name': 'volume-a34e5b41-b9fc-4c5a-967e-b3e601dde931', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b7b2b7ac-2321-4cf3-b2ed-4877b893af63', 'attached_at': '', 'detached_at': '', 'volume_id': 'a34e5b41-b9fc-4c5a-967e-b3e601dde931', 'serial': 'a34e5b41-b9fc-4c5a-967e-b3e601dde931'} {{(pid=63531) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1054.937359] env[63531]: DEBUG oslo_vmware.api [None req-09413a86-09cb-400a-aa04-8dbb3cff6da3 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118531, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.943309] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1054.955891] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-584426ce-5cc1-4420-87bd-4a33f7084720 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] [instance: 3e7b8a01-d930-43cf-9432-0dabc68da20e] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1054.956137] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-584426ce-5cc1-4420-87bd-4a33f7084720 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] [instance: 3e7b8a01-d930-43cf-9432-0dabc68da20e] Deleting contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1054.956330] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-584426ce-5cc1-4420-87bd-4a33f7084720 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Deleting the datastore file [datastore2] 3e7b8a01-d930-43cf-9432-0dabc68da20e {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1054.956605] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-972a7b6c-2edc-4d69-90ee-177fbd326479 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.963516] env[63531]: DEBUG oslo_vmware.api [None req-584426ce-5cc1-4420-87bd-4a33f7084720 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Waiting for the task: (returnval){ [ 1054.963516] env[63531]: value = "task-1118533" [ 1054.963516] env[63531]: _type = "Task" [ 1054.963516] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.973528] env[63531]: DEBUG oslo_vmware.api [None req-584426ce-5cc1-4420-87bd-4a33f7084720 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Task: {'id': task-1118533, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.040170] env[63531]: DEBUG nova.compute.manager [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 8f903d83-adbd-4f71-a6f1-037807746248] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1055.049829] env[63531]: INFO nova.compute.resource_tracker [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Updating resource usage from migration b6672930-332c-4293-8ab4-8ad450b70f9d [ 1055.149789] env[63531]: DEBUG nova.network.neutron [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 8f903d83-adbd-4f71-a6f1-037807746248] Successfully created port: 0319492b-6961-4fcd-92f1-ddba0968522d {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1055.274812] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7735738c-74c9-4d08-95f0-d401df8b1e5e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.283225] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e0768af-590c-43be-b6f5-078e0264db9b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.314740] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-010d6185-e345-4d66-95be-9defeb3a28b4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.323353] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47fae8eb-c32b-4c93-b920-3e91cb960561 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.338135] env[63531]: DEBUG nova.compute.provider_tree [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1055.428281] env[63531]: DEBUG oslo_vmware.api [None req-09413a86-09cb-400a-aa04-8dbb3cff6da3 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118531, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.474028] env[63531]: DEBUG oslo_vmware.api [None req-584426ce-5cc1-4420-87bd-4a33f7084720 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Task: {'id': task-1118533, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.157652} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.474392] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-584426ce-5cc1-4420-87bd-4a33f7084720 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1055.474646] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-584426ce-5cc1-4420-87bd-4a33f7084720 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] [instance: 3e7b8a01-d930-43cf-9432-0dabc68da20e] Deleted contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1055.474887] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-584426ce-5cc1-4420-87bd-4a33f7084720 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] [instance: 3e7b8a01-d930-43cf-9432-0dabc68da20e] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1055.475159] env[63531]: INFO nova.compute.manager [None req-584426ce-5cc1-4420-87bd-4a33f7084720 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] [instance: 3e7b8a01-d930-43cf-9432-0dabc68da20e] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1055.475491] env[63531]: DEBUG oslo.service.loopingcall [None req-584426ce-5cc1-4420-87bd-4a33f7084720 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1055.475753] env[63531]: DEBUG nova.compute.manager [-] [instance: 3e7b8a01-d930-43cf-9432-0dabc68da20e] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1055.475893] env[63531]: DEBUG nova.network.neutron [-] [instance: 3e7b8a01-d930-43cf-9432-0dabc68da20e] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1055.491797] env[63531]: DEBUG nova.network.neutron [-] [instance: 3e7b8a01-d930-43cf-9432-0dabc68da20e] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1055.841388] env[63531]: DEBUG nova.scheduler.client.report [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1055.926428] env[63531]: DEBUG oslo_vmware.api [None req-09413a86-09cb-400a-aa04-8dbb3cff6da3 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118531, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.961303] env[63531]: DEBUG nova.objects.instance [None req-11bea7a3-32cd-4f38-8cbd-f4972d5dad3e tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lazy-loading 'flavor' on Instance uuid b7b2b7ac-2321-4cf3-b2ed-4877b893af63 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1055.993861] env[63531]: DEBUG nova.network.neutron [-] [instance: 3e7b8a01-d930-43cf-9432-0dabc68da20e] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1056.050521] env[63531]: DEBUG nova.compute.manager [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 8f903d83-adbd-4f71-a6f1-037807746248] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1056.093157] env[63531]: DEBUG nova.virt.hardware [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1056.093422] env[63531]: DEBUG nova.virt.hardware [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1056.093584] env[63531]: DEBUG nova.virt.hardware [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1056.093770] env[63531]: DEBUG nova.virt.hardware [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1056.093924] env[63531]: DEBUG nova.virt.hardware [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1056.094136] env[63531]: DEBUG nova.virt.hardware [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1056.094368] env[63531]: DEBUG nova.virt.hardware [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1056.094536] env[63531]: DEBUG nova.virt.hardware [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1056.094710] env[63531]: DEBUG nova.virt.hardware [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1056.094880] env[63531]: DEBUG nova.virt.hardware [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1056.095076] env[63531]: DEBUG nova.virt.hardware [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1056.096212] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00ae8979-6297-44c7-bab5-30e2823f25f2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.105401] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6987140-e41e-4c29-801b-f80cc6232655 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.348086] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.313s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1056.348086] env[63531]: INFO nova.compute.manager [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Migrating [ 1056.352980] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0e8b28d-c0dc-41b1-b98b-a77c5c3f4e22 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.781s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.353223] env[63531]: DEBUG nova.objects.instance [None req-c0e8b28d-c0dc-41b1-b98b-a77c5c3f4e22 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Lazy-loading 'resources' on Instance uuid 879924d4-b465-4102-a0e3-c7b2be7ef08b {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1056.433705] env[63531]: DEBUG oslo_vmware.api [None req-09413a86-09cb-400a-aa04-8dbb3cff6da3 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118531, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.466306] env[63531]: DEBUG oslo_concurrency.lockutils [None req-11bea7a3-32cd-4f38-8cbd-f4972d5dad3e tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lock "b7b2b7ac-2321-4cf3-b2ed-4877b893af63" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.382s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1056.496698] env[63531]: INFO nova.compute.manager [-] [instance: 3e7b8a01-d930-43cf-9432-0dabc68da20e] Took 1.02 seconds to deallocate network for instance. [ 1056.565354] env[63531]: DEBUG nova.compute.manager [req-d0d3ea96-9c32-4567-ad0a-7ebf6605eb9e req-46277d85-95f0-41cd-8a27-c9d60002d8f3 service nova] [instance: 8f903d83-adbd-4f71-a6f1-037807746248] Received event network-vif-plugged-0319492b-6961-4fcd-92f1-ddba0968522d {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1056.565587] env[63531]: DEBUG oslo_concurrency.lockutils [req-d0d3ea96-9c32-4567-ad0a-7ebf6605eb9e req-46277d85-95f0-41cd-8a27-c9d60002d8f3 service nova] Acquiring lock "8f903d83-adbd-4f71-a6f1-037807746248-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1056.565804] env[63531]: DEBUG oslo_concurrency.lockutils [req-d0d3ea96-9c32-4567-ad0a-7ebf6605eb9e req-46277d85-95f0-41cd-8a27-c9d60002d8f3 service nova] Lock "8f903d83-adbd-4f71-a6f1-037807746248-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.565982] env[63531]: DEBUG oslo_concurrency.lockutils [req-d0d3ea96-9c32-4567-ad0a-7ebf6605eb9e req-46277d85-95f0-41cd-8a27-c9d60002d8f3 service nova] Lock "8f903d83-adbd-4f71-a6f1-037807746248-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1056.566178] env[63531]: DEBUG nova.compute.manager [req-d0d3ea96-9c32-4567-ad0a-7ebf6605eb9e req-46277d85-95f0-41cd-8a27-c9d60002d8f3 service nova] [instance: 8f903d83-adbd-4f71-a6f1-037807746248] No waiting events found dispatching network-vif-plugged-0319492b-6961-4fcd-92f1-ddba0968522d {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1056.566354] env[63531]: WARNING nova.compute.manager [req-d0d3ea96-9c32-4567-ad0a-7ebf6605eb9e req-46277d85-95f0-41cd-8a27-c9d60002d8f3 service nova] [instance: 8f903d83-adbd-4f71-a6f1-037807746248] Received unexpected event network-vif-plugged-0319492b-6961-4fcd-92f1-ddba0968522d for instance with vm_state building and task_state spawning. [ 1056.589644] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffd11e88-9b21-4a0b-8c36-80583899f648 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.598465] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8269cb58-4a9e-4f58-af48-37ddeea82dad {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.628704] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3eb1336-127e-4cf1-b871-15271631fb79 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.637672] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad4addbc-5ded-412f-a2ae-3efa8bd61f16 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.650706] env[63531]: DEBUG nova.compute.provider_tree [None req-c0e8b28d-c0dc-41b1-b98b-a77c5c3f4e22 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1056.665215] env[63531]: DEBUG nova.network.neutron [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 8f903d83-adbd-4f71-a6f1-037807746248] Successfully updated port: 0319492b-6961-4fcd-92f1-ddba0968522d {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1056.867797] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquiring lock "refresh_cache-18057005-1ff1-4e3d-bccc-8c89755d6db7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1056.867797] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquired lock "refresh_cache-18057005-1ff1-4e3d-bccc-8c89755d6db7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1056.867957] env[63531]: DEBUG nova.network.neutron [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1056.933434] env[63531]: DEBUG oslo_vmware.api [None req-09413a86-09cb-400a-aa04-8dbb3cff6da3 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118531, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.004567] env[63531]: DEBUG oslo_concurrency.lockutils [None req-584426ce-5cc1-4420-87bd-4a33f7084720 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1057.153302] env[63531]: DEBUG nova.scheduler.client.report [None req-c0e8b28d-c0dc-41b1-b98b-a77c5c3f4e22 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1057.168438] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "refresh_cache-8f903d83-adbd-4f71-a6f1-037807746248" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1057.168588] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquired lock "refresh_cache-8f903d83-adbd-4f71-a6f1-037807746248" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1057.168800] env[63531]: DEBUG nova.network.neutron [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 8f903d83-adbd-4f71-a6f1-037807746248] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1057.423886] env[63531]: DEBUG oslo_concurrency.lockutils [None req-455f3e56-7a45-4cca-bf4f-e08d6600b671 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Acquiring lock "b7b2b7ac-2321-4cf3-b2ed-4877b893af63" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1057.424134] env[63531]: DEBUG oslo_concurrency.lockutils [None req-455f3e56-7a45-4cca-bf4f-e08d6600b671 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lock "b7b2b7ac-2321-4cf3-b2ed-4877b893af63" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1057.428557] env[63531]: DEBUG oslo_vmware.api [None req-09413a86-09cb-400a-aa04-8dbb3cff6da3 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118531, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.567795] env[63531]: DEBUG nova.network.neutron [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Updating instance_info_cache with network_info: [{"id": "af2485f6-5ce2-4a83-9d95-7bea85cebff2", "address": "fa:16:3e:f0:7d:92", "network": {"id": "c4f4c994-36ab-42ab-94f8-78f29cbafa50", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-994507698-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "57e9bf4ee0f1463fa67b0f8deff30fdc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "767a3a48-41d4-4a0c-961d-0024837f63bd", "external-id": "nsx-vlan-transportzone-141", "segmentation_id": 141, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaf2485f6-5c", "ovs_interfaceid": "af2485f6-5ce2-4a83-9d95-7bea85cebff2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1057.657623] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0e8b28d-c0dc-41b1-b98b-a77c5c3f4e22 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.305s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.660028] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.323s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1057.661452] env[63531]: INFO nova.compute.claims [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 85058146-ba4c-4e9a-bbe7-048ed56167ab] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1057.681411] env[63531]: INFO nova.scheduler.client.report [None req-c0e8b28d-c0dc-41b1-b98b-a77c5c3f4e22 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Deleted allocations for instance 879924d4-b465-4102-a0e3-c7b2be7ef08b [ 1057.699644] env[63531]: DEBUG nova.network.neutron [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 8f903d83-adbd-4f71-a6f1-037807746248] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1057.823919] env[63531]: DEBUG nova.network.neutron [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 8f903d83-adbd-4f71-a6f1-037807746248] Updating instance_info_cache with network_info: [{"id": "0319492b-6961-4fcd-92f1-ddba0968522d", "address": "fa:16:3e:91:78:dd", "network": {"id": "ea2d47d3-a502-4afd-9bc5-ee4a37017901", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1361282648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94a95667c00947ea9183a6307c569c90", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6c45fd-e930-495a-9cb7-df84eda443b1", "external-id": "nsx-vlan-transportzone-407", "segmentation_id": 407, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0319492b-69", "ovs_interfaceid": "0319492b-6961-4fcd-92f1-ddba0968522d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1057.928931] env[63531]: DEBUG oslo_vmware.api [None req-09413a86-09cb-400a-aa04-8dbb3cff6da3 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118531, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.930008] env[63531]: DEBUG nova.compute.utils [None req-455f3e56-7a45-4cca-bf4f-e08d6600b671 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1058.070361] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Releasing lock "refresh_cache-18057005-1ff1-4e3d-bccc-8c89755d6db7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1058.188185] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c0e8b28d-c0dc-41b1-b98b-a77c5c3f4e22 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Lock "879924d4-b465-4102-a0e3-c7b2be7ef08b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.583s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.325604] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Releasing lock "refresh_cache-8f903d83-adbd-4f71-a6f1-037807746248" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1058.325937] env[63531]: DEBUG nova.compute.manager [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 8f903d83-adbd-4f71-a6f1-037807746248] Instance network_info: |[{"id": "0319492b-6961-4fcd-92f1-ddba0968522d", "address": "fa:16:3e:91:78:dd", "network": {"id": "ea2d47d3-a502-4afd-9bc5-ee4a37017901", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1361282648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94a95667c00947ea9183a6307c569c90", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6c45fd-e930-495a-9cb7-df84eda443b1", "external-id": "nsx-vlan-transportzone-407", "segmentation_id": 407, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0319492b-69", "ovs_interfaceid": "0319492b-6961-4fcd-92f1-ddba0968522d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1058.326404] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 8f903d83-adbd-4f71-a6f1-037807746248] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:91:78:dd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2b6c45fd-e930-495a-9cb7-df84eda443b1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0319492b-6961-4fcd-92f1-ddba0968522d', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1058.334077] env[63531]: DEBUG oslo.service.loopingcall [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1058.334638] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8f903d83-adbd-4f71-a6f1-037807746248] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1058.334882] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e0f4bffe-0356-40c1-8504-b86c2e58749a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.356126] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1058.356126] env[63531]: value = "task-1118536" [ 1058.356126] env[63531]: _type = "Task" [ 1058.356126] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.366633] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118536, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.430171] env[63531]: DEBUG oslo_vmware.api [None req-09413a86-09cb-400a-aa04-8dbb3cff6da3 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118531, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.433993] env[63531]: DEBUG oslo_concurrency.lockutils [None req-455f3e56-7a45-4cca-bf4f-e08d6600b671 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lock "b7b2b7ac-2321-4cf3-b2ed-4877b893af63" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.010s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.610168] env[63531]: DEBUG nova.compute.manager [req-87051672-248a-4566-9264-dc971924bdd8 req-f924501d-ed9d-4a0c-b7d7-712c5b7beb9f service nova] [instance: 8f903d83-adbd-4f71-a6f1-037807746248] Received event network-changed-0319492b-6961-4fcd-92f1-ddba0968522d {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1058.610430] env[63531]: DEBUG nova.compute.manager [req-87051672-248a-4566-9264-dc971924bdd8 req-f924501d-ed9d-4a0c-b7d7-712c5b7beb9f service nova] [instance: 8f903d83-adbd-4f71-a6f1-037807746248] Refreshing instance network info cache due to event network-changed-0319492b-6961-4fcd-92f1-ddba0968522d. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1058.610683] env[63531]: DEBUG oslo_concurrency.lockutils [req-87051672-248a-4566-9264-dc971924bdd8 req-f924501d-ed9d-4a0c-b7d7-712c5b7beb9f service nova] Acquiring lock "refresh_cache-8f903d83-adbd-4f71-a6f1-037807746248" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1058.610893] env[63531]: DEBUG oslo_concurrency.lockutils [req-87051672-248a-4566-9264-dc971924bdd8 req-f924501d-ed9d-4a0c-b7d7-712c5b7beb9f service nova] Acquired lock "refresh_cache-8f903d83-adbd-4f71-a6f1-037807746248" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.611091] env[63531]: DEBUG nova.network.neutron [req-87051672-248a-4566-9264-dc971924bdd8 req-f924501d-ed9d-4a0c-b7d7-712c5b7beb9f service nova] [instance: 8f903d83-adbd-4f71-a6f1-037807746248] Refreshing network info cache for port 0319492b-6961-4fcd-92f1-ddba0968522d {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1058.837882] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9dbfc3a-92a4-4ed1-8280-e91bcaafd834 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.845887] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8715a5d-f5ea-48ab-aa94-c46ecd11765c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.879815] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14015753-0461-4738-a782-fc4b74193b5b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.889526] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118536, 'name': CreateVM_Task, 'duration_secs': 0.320422} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.890710] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad448639-aa3e-4e64-884a-027c2b69cd6c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.894130] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8f903d83-adbd-4f71-a6f1-037807746248] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1058.894806] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1058.894976] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.895314] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1058.895857] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d981ce9b-f9de-47d9-a29b-5611cf80968f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.907195] env[63531]: DEBUG nova.compute.provider_tree [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1058.909328] env[63531]: DEBUG oslo_vmware.api [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 1058.909328] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]522bf0d8-e4c5-0fa3-98b7-7fd1e971a728" [ 1058.909328] env[63531]: _type = "Task" [ 1058.909328] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.919370] env[63531]: DEBUG oslo_vmware.api [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]522bf0d8-e4c5-0fa3-98b7-7fd1e971a728, 'name': SearchDatastore_Task, 'duration_secs': 0.009856} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.919643] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1058.919880] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 8f903d83-adbd-4f71-a6f1-037807746248] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1058.921101] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1058.921101] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.921101] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1058.921101] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-70383baf-1e71-4e3d-8730-6d3f96a1e052 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.934632] env[63531]: DEBUG oslo_vmware.api [None req-09413a86-09cb-400a-aa04-8dbb3cff6da3 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118531, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.935682] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1058.935788] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1058.936565] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c9366f9c-7ee2-4703-9ec7-493b254445c8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.942477] env[63531]: DEBUG oslo_vmware.api [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 1058.942477] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52411389-298a-7aa2-c3fe-b9bf603a59f5" [ 1058.942477] env[63531]: _type = "Task" [ 1058.942477] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.950830] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5dff6cb1-9c4d-4e3f-bec7-275c4d5265b6 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquiring lock "4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1058.951120] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5dff6cb1-9c4d-4e3f-bec7-275c4d5265b6 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Lock "4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1058.951337] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5dff6cb1-9c4d-4e3f-bec7-275c4d5265b6 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquiring lock "4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1058.951525] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5dff6cb1-9c4d-4e3f-bec7-275c4d5265b6 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Lock "4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1058.951697] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5dff6cb1-9c4d-4e3f-bec7-275c4d5265b6 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Lock "4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.953307] env[63531]: DEBUG oslo_vmware.api [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52411389-298a-7aa2-c3fe-b9bf603a59f5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.954141] env[63531]: INFO nova.compute.manager [None req-5dff6cb1-9c4d-4e3f-bec7-275c4d5265b6 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Terminating instance [ 1058.955932] env[63531]: DEBUG nova.compute.manager [None req-5dff6cb1-9c4d-4e3f-bec7-275c4d5265b6 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1058.956166] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5dff6cb1-9c4d-4e3f-bec7-275c4d5265b6 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1058.957188] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-748c03e5-867a-4211-a10d-c133d4d73567 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.964266] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-5dff6cb1-9c4d-4e3f-bec7-275c4d5265b6 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1058.964496] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-afdd2a2d-bbcd-4a96-b9c8-819a9ef7c5ba {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.971321] env[63531]: DEBUG oslo_vmware.api [None req-5dff6cb1-9c4d-4e3f-bec7-275c4d5265b6 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 1058.971321] env[63531]: value = "task-1118537" [ 1058.971321] env[63531]: _type = "Task" [ 1058.971321] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.981201] env[63531]: DEBUG oslo_vmware.api [None req-5dff6cb1-9c4d-4e3f-bec7-275c4d5265b6 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118537, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.387195] env[63531]: DEBUG nova.network.neutron [req-87051672-248a-4566-9264-dc971924bdd8 req-f924501d-ed9d-4a0c-b7d7-712c5b7beb9f service nova] [instance: 8f903d83-adbd-4f71-a6f1-037807746248] Updated VIF entry in instance network info cache for port 0319492b-6961-4fcd-92f1-ddba0968522d. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1059.387584] env[63531]: DEBUG nova.network.neutron [req-87051672-248a-4566-9264-dc971924bdd8 req-f924501d-ed9d-4a0c-b7d7-712c5b7beb9f service nova] [instance: 8f903d83-adbd-4f71-a6f1-037807746248] Updating instance_info_cache with network_info: [{"id": "0319492b-6961-4fcd-92f1-ddba0968522d", "address": "fa:16:3e:91:78:dd", "network": {"id": "ea2d47d3-a502-4afd-9bc5-ee4a37017901", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1361282648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94a95667c00947ea9183a6307c569c90", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6c45fd-e930-495a-9cb7-df84eda443b1", "external-id": "nsx-vlan-transportzone-407", "segmentation_id": 407, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0319492b-69", "ovs_interfaceid": "0319492b-6961-4fcd-92f1-ddba0968522d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1059.411318] env[63531]: DEBUG nova.scheduler.client.report [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1059.430637] env[63531]: DEBUG oslo_vmware.api [None req-09413a86-09cb-400a-aa04-8dbb3cff6da3 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118531, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.453085] env[63531]: DEBUG oslo_vmware.api [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52411389-298a-7aa2-c3fe-b9bf603a59f5, 'name': SearchDatastore_Task, 'duration_secs': 0.010329} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.453906] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-33a952cc-d846-4dac-9460-624a7d5647fb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.458909] env[63531]: DEBUG oslo_vmware.api [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 1059.458909] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52b677de-dbb1-bedb-5583-ca423941287f" [ 1059.458909] env[63531]: _type = "Task" [ 1059.458909] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.467452] env[63531]: DEBUG oslo_vmware.api [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52b677de-dbb1-bedb-5583-ca423941287f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.478793] env[63531]: DEBUG oslo_vmware.api [None req-5dff6cb1-9c4d-4e3f-bec7-275c4d5265b6 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118537, 'name': PowerOffVM_Task, 'duration_secs': 0.216802} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.479063] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-5dff6cb1-9c4d-4e3f-bec7-275c4d5265b6 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1059.479244] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5dff6cb1-9c4d-4e3f-bec7-275c4d5265b6 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1059.479506] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1a7d53f0-706e-43e6-ad36-599d9fb4907a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.498993] env[63531]: DEBUG oslo_concurrency.lockutils [None req-455f3e56-7a45-4cca-bf4f-e08d6600b671 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Acquiring lock "b7b2b7ac-2321-4cf3-b2ed-4877b893af63" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1059.499242] env[63531]: DEBUG oslo_concurrency.lockutils [None req-455f3e56-7a45-4cca-bf4f-e08d6600b671 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lock "b7b2b7ac-2321-4cf3-b2ed-4877b893af63" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.499472] env[63531]: INFO nova.compute.manager [None req-455f3e56-7a45-4cca-bf4f-e08d6600b671 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Attaching volume 13d7c941-f906-4020-b545-f2bf1bf8a7a0 to /dev/sdc [ 1059.537092] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aea89b49-753b-430f-8a96-76b1a5a15d79 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.544497] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b94a0d7-7fa4-4bb0-83d5-cc72d9d57841 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.550272] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5dff6cb1-9c4d-4e3f-bec7-275c4d5265b6 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1059.550502] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5dff6cb1-9c4d-4e3f-bec7-275c4d5265b6 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1059.550731] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-5dff6cb1-9c4d-4e3f-bec7-275c4d5265b6 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Deleting the datastore file [datastore1] 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1059.550984] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-375a173d-5b76-4abb-83c7-f9f949ec665c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.560937] env[63531]: DEBUG nova.virt.block_device [None req-455f3e56-7a45-4cca-bf4f-e08d6600b671 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Updating existing volume attachment record: 3ecc6c5c-e3a6-4ab7-a0b1-ba48dcc1e436 {{(pid=63531) _volume_attach /opt/stack/nova/nova/virt/block_device.py:679}} [ 1059.564286] env[63531]: DEBUG oslo_vmware.api [None req-5dff6cb1-9c4d-4e3f-bec7-275c4d5265b6 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for the task: (returnval){ [ 1059.564286] env[63531]: value = "task-1118540" [ 1059.564286] env[63531]: _type = "Task" [ 1059.564286] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.572076] env[63531]: DEBUG oslo_vmware.api [None req-5dff6cb1-9c4d-4e3f-bec7-275c4d5265b6 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118540, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.586973] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-800616c2-7c9f-4df2-a1d4-31042145401b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.607714] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Updating instance '18057005-1ff1-4e3d-bccc-8c89755d6db7' progress to 0 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1059.891072] env[63531]: DEBUG oslo_concurrency.lockutils [req-87051672-248a-4566-9264-dc971924bdd8 req-f924501d-ed9d-4a0c-b7d7-712c5b7beb9f service nova] Releasing lock "refresh_cache-8f903d83-adbd-4f71-a6f1-037807746248" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1059.915935] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.256s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1059.916615] env[63531]: DEBUG nova.compute.manager [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 85058146-ba4c-4e9a-bbe7-048ed56167ab] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1059.919397] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b92663e6-8900-4f7f-bc86-533aa67ecc8f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.371s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.919607] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b92663e6-8900-4f7f-bc86-533aa67ecc8f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1059.922191] env[63531]: DEBUG oslo_concurrency.lockutils [None req-aa491dbe-51b5-432e-ab71-8cdd1877a2f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.166s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.922415] env[63531]: DEBUG nova.objects.instance [None req-aa491dbe-51b5-432e-ab71-8cdd1877a2f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lazy-loading 'resources' on Instance uuid 97d3f108-e299-4ef1-84ca-532cd64905d5 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1059.933850] env[63531]: DEBUG oslo_vmware.api [None req-09413a86-09cb-400a-aa04-8dbb3cff6da3 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118531, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.946037] env[63531]: INFO nova.scheduler.client.report [None req-b92663e6-8900-4f7f-bc86-533aa67ecc8f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Deleted allocations for instance 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7 [ 1059.969572] env[63531]: DEBUG oslo_vmware.api [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52b677de-dbb1-bedb-5583-ca423941287f, 'name': SearchDatastore_Task, 'duration_secs': 0.009504} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.969910] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1059.970661] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] 8f903d83-adbd-4f71-a6f1-037807746248/8f903d83-adbd-4f71-a6f1-037807746248.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1059.970661] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5739f5ad-c3e3-4b21-b440-3f676564e44e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.984566] env[63531]: DEBUG oslo_vmware.api [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 1059.984566] env[63531]: value = "task-1118542" [ 1059.984566] env[63531]: _type = "Task" [ 1059.984566] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.995837] env[63531]: DEBUG oslo_vmware.api [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118542, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.075694] env[63531]: DEBUG oslo_vmware.api [None req-5dff6cb1-9c4d-4e3f-bec7-275c4d5265b6 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Task: {'id': task-1118540, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.152702} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.076011] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-5dff6cb1-9c4d-4e3f-bec7-275c4d5265b6 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1060.076249] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5dff6cb1-9c4d-4e3f-bec7-275c4d5265b6 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1060.076477] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5dff6cb1-9c4d-4e3f-bec7-275c4d5265b6 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1060.076732] env[63531]: INFO nova.compute.manager [None req-5dff6cb1-9c4d-4e3f-bec7-275c4d5265b6 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1060.077014] env[63531]: DEBUG oslo.service.loopingcall [None req-5dff6cb1-9c4d-4e3f-bec7-275c4d5265b6 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1060.077258] env[63531]: DEBUG nova.compute.manager [-] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1060.077371] env[63531]: DEBUG nova.network.neutron [-] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1060.115787] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1060.116101] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-871f091b-b30d-4d39-a7e0-5ba4a6130187 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.124210] env[63531]: DEBUG oslo_vmware.api [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 1060.124210] env[63531]: value = "task-1118543" [ 1060.124210] env[63531]: _type = "Task" [ 1060.124210] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.135516] env[63531]: DEBUG oslo_vmware.api [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118543, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.425089] env[63531]: DEBUG nova.compute.utils [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1060.430209] env[63531]: DEBUG nova.compute.manager [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 85058146-ba4c-4e9a-bbe7-048ed56167ab] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1060.430209] env[63531]: DEBUG nova.network.neutron [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 85058146-ba4c-4e9a-bbe7-048ed56167ab] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1060.443272] env[63531]: DEBUG oslo_vmware.api [None req-09413a86-09cb-400a-aa04-8dbb3cff6da3 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118531, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.456263] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b92663e6-8900-4f7f-bc86-533aa67ecc8f tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.023s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1060.485827] env[63531]: DEBUG nova.policy [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '049cb89a769f41bd91d0e62784635cbd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c2e99349303e413b9f7896d449cb7dff', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 1060.501129] env[63531]: DEBUG oslo_vmware.api [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118542, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.510298} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.501491] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] 8f903d83-adbd-4f71-a6f1-037807746248/8f903d83-adbd-4f71-a6f1-037807746248.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1060.501806] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 8f903d83-adbd-4f71-a6f1-037807746248] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1060.502068] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-388dc6be-e843-464f-97d8-88bc85c5b64c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.511042] env[63531]: DEBUG oslo_vmware.api [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 1060.511042] env[63531]: value = "task-1118544" [ 1060.511042] env[63531]: _type = "Task" [ 1060.511042] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.243242] env[63531]: DEBUG nova.network.neutron [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 85058146-ba4c-4e9a-bbe7-048ed56167ab] Successfully created port: 9216e469-ea4e-4e8a-9f74-e9310da25110 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1061.244582] env[63531]: DEBUG nova.network.neutron [-] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1061.245990] env[63531]: DEBUG nova.compute.manager [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 85058146-ba4c-4e9a-bbe7-048ed56167ab] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1061.251369] env[63531]: DEBUG nova.compute.manager [req-1ad24ab9-3970-4743-883c-24bac7b682d6 req-5170e357-bdf2-470f-812a-2b70c3867669 service nova] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Received event network-vif-deleted-e30711c3-e609-434b-82da-a58252d3678d {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1061.251369] env[63531]: INFO nova.compute.manager [req-1ad24ab9-3970-4743-883c-24bac7b682d6 req-5170e357-bdf2-470f-812a-2b70c3867669 service nova] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Neutron deleted interface e30711c3-e609-434b-82da-a58252d3678d; detaching it from the instance and deleting it from the info cache [ 1061.251369] env[63531]: DEBUG nova.network.neutron [req-1ad24ab9-3970-4743-883c-24bac7b682d6 req-5170e357-bdf2-470f-812a-2b70c3867669 service nova] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1061.263746] env[63531]: DEBUG oslo_vmware.api [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118544, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074639} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.270244] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 8f903d83-adbd-4f71-a6f1-037807746248] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1061.270590] env[63531]: DEBUG oslo_vmware.api [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118543, 'name': PowerOffVM_Task} progress is 100%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.271449] env[63531]: DEBUG oslo_vmware.api [None req-09413a86-09cb-400a-aa04-8dbb3cff6da3 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118531, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.271520] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73a71ce7-da15-4118-92b0-6c1ac2ea1c2b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.298288] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 8f903d83-adbd-4f71-a6f1-037807746248] Reconfiguring VM instance instance-00000067 to attach disk [datastore2] 8f903d83-adbd-4f71-a6f1-037807746248/8f903d83-adbd-4f71-a6f1-037807746248.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1061.299563] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d1e04e53-8b4b-46b5-9d1a-e93c670a4445 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.327334] env[63531]: DEBUG oslo_vmware.api [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 1061.327334] env[63531]: value = "task-1118546" [ 1061.327334] env[63531]: _type = "Task" [ 1061.327334] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.337688] env[63531]: DEBUG oslo_vmware.api [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118546, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.460708] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d01c6e75-dbc7-45c0-9f6d-95d46d70c932 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.469192] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a460b29-0111-4c0b-8e1d-acc59be68e9a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.502306] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d4bb853-75e3-4cfe-9d5e-e6e45b926890 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.510532] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9331ba2c-7ec8-432e-bc2a-06306ae46c7a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.524567] env[63531]: DEBUG nova.compute.provider_tree [None req-aa491dbe-51b5-432e-ab71-8cdd1877a2f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1061.756934] env[63531]: INFO nova.compute.manager [-] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Took 1.68 seconds to deallocate network for instance. [ 1061.765161] env[63531]: DEBUG oslo_vmware.api [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118543, 'name': PowerOffVM_Task, 'duration_secs': 1.305689} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.772374] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-351b77f3-45f4-4d19-a790-0d3c3805b91e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.778420] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1061.778657] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Updating instance '18057005-1ff1-4e3d-bccc-8c89755d6db7' progress to 17 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1061.789654] env[63531]: DEBUG oslo_vmware.api [None req-09413a86-09cb-400a-aa04-8dbb3cff6da3 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118531, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.803027] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94dd584e-dbde-4dce-a5a7-985551f4fa3f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.843934] env[63531]: DEBUG nova.compute.manager [req-1ad24ab9-3970-4743-883c-24bac7b682d6 req-5170e357-bdf2-470f-812a-2b70c3867669 service nova] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Detach interface failed, port_id=e30711c3-e609-434b-82da-a58252d3678d, reason: Instance 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df could not be found. {{(pid=63531) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1061.850655] env[63531]: DEBUG oslo_vmware.api [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118546, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.028017] env[63531]: DEBUG nova.scheduler.client.report [None req-aa491dbe-51b5-432e-ab71-8cdd1877a2f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1062.261557] env[63531]: DEBUG nova.compute.manager [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 85058146-ba4c-4e9a-bbe7-048ed56167ab] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1062.263661] env[63531]: DEBUG oslo_vmware.api [None req-09413a86-09cb-400a-aa04-8dbb3cff6da3 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118531, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.273139] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5dff6cb1-9c4d-4e3f-bec7-275c4d5265b6 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1062.286722] env[63531]: DEBUG nova.virt.hardware [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1062.287068] env[63531]: DEBUG nova.virt.hardware [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1062.287238] env[63531]: DEBUG nova.virt.hardware [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1062.287428] env[63531]: DEBUG nova.virt.hardware [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1062.287573] env[63531]: DEBUG nova.virt.hardware [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1062.287742] env[63531]: DEBUG nova.virt.hardware [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1062.287946] env[63531]: DEBUG nova.virt.hardware [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1062.288149] env[63531]: DEBUG nova.virt.hardware [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1062.288332] env[63531]: DEBUG nova.virt.hardware [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1062.288591] env[63531]: DEBUG nova.virt.hardware [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1062.288883] env[63531]: DEBUG nova.virt.hardware [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1062.294383] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ec94ec97-9c02-4b91-a16e-11bd75769165 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.305959] env[63531]: DEBUG nova.virt.hardware [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1062.306120] env[63531]: DEBUG nova.virt.hardware [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1062.306256] env[63531]: DEBUG nova.virt.hardware [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1062.306444] env[63531]: DEBUG nova.virt.hardware [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1062.306593] env[63531]: DEBUG nova.virt.hardware [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1062.306785] env[63531]: DEBUG nova.virt.hardware [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1062.306932] env[63531]: DEBUG nova.virt.hardware [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1062.307104] env[63531]: DEBUG nova.virt.hardware [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1062.307271] env[63531]: DEBUG nova.virt.hardware [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1062.307434] env[63531]: DEBUG nova.virt.hardware [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1062.307609] env[63531]: DEBUG nova.virt.hardware [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1062.308936] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-280eb1d7-ce50-4e7f-b3af-dbe22990abea {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.317418] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0adae661-7c2a-49b6-96d1-37e0761c77b9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.322129] env[63531]: DEBUG oslo_vmware.api [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 1062.322129] env[63531]: value = "task-1118548" [ 1062.322129] env[63531]: _type = "Task" [ 1062.322129] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.339223] env[63531]: DEBUG oslo_vmware.api [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118548, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.342496] env[63531]: DEBUG oslo_vmware.api [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118546, 'name': ReconfigVM_Task, 'duration_secs': 0.587587} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.342788] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 8f903d83-adbd-4f71-a6f1-037807746248] Reconfigured VM instance instance-00000067 to attach disk [datastore2] 8f903d83-adbd-4f71-a6f1-037807746248/8f903d83-adbd-4f71-a6f1-037807746248.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1062.343427] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f10bf5e4-9424-4892-9b8c-793edc89beb9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.349491] env[63531]: DEBUG oslo_vmware.api [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 1062.349491] env[63531]: value = "task-1118549" [ 1062.349491] env[63531]: _type = "Task" [ 1062.349491] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.357607] env[63531]: DEBUG oslo_vmware.api [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118549, 'name': Rename_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.533989] env[63531]: DEBUG oslo_concurrency.lockutils [None req-aa491dbe-51b5-432e-ab71-8cdd1877a2f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.612s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1062.536412] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.593s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1062.537962] env[63531]: INFO nova.compute.claims [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1062.558755] env[63531]: INFO nova.scheduler.client.report [None req-aa491dbe-51b5-432e-ab71-8cdd1877a2f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Deleted allocations for instance 97d3f108-e299-4ef1-84ca-532cd64905d5 [ 1062.760604] env[63531]: DEBUG oslo_vmware.api [None req-09413a86-09cb-400a-aa04-8dbb3cff6da3 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118531, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.832392] env[63531]: DEBUG oslo_vmware.api [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118548, 'name': ReconfigVM_Task, 'duration_secs': 0.173655} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.832717] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Updating instance '18057005-1ff1-4e3d-bccc-8c89755d6db7' progress to 33 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1062.860227] env[63531]: DEBUG oslo_vmware.api [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118549, 'name': Rename_Task, 'duration_secs': 0.162944} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.860578] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 8f903d83-adbd-4f71-a6f1-037807746248] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1062.860872] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1b943730-a749-4712-8edf-797ef382ea4e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.869214] env[63531]: DEBUG oslo_vmware.api [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 1062.869214] env[63531]: value = "task-1118550" [ 1062.869214] env[63531]: _type = "Task" [ 1062.869214] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.880225] env[63531]: DEBUG oslo_vmware.api [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118550, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.018522] env[63531]: DEBUG nova.compute.manager [req-d029a0d2-c695-4ccb-851a-f090a0ce7dfd req-6ebd944e-783d-4fcf-91b4-da5652e69d62 service nova] [instance: 85058146-ba4c-4e9a-bbe7-048ed56167ab] Received event network-vif-plugged-9216e469-ea4e-4e8a-9f74-e9310da25110 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1063.018811] env[63531]: DEBUG oslo_concurrency.lockutils [req-d029a0d2-c695-4ccb-851a-f090a0ce7dfd req-6ebd944e-783d-4fcf-91b4-da5652e69d62 service nova] Acquiring lock "85058146-ba4c-4e9a-bbe7-048ed56167ab-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.019115] env[63531]: DEBUG oslo_concurrency.lockutils [req-d029a0d2-c695-4ccb-851a-f090a0ce7dfd req-6ebd944e-783d-4fcf-91b4-da5652e69d62 service nova] Lock "85058146-ba4c-4e9a-bbe7-048ed56167ab-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.019405] env[63531]: DEBUG oslo_concurrency.lockutils [req-d029a0d2-c695-4ccb-851a-f090a0ce7dfd req-6ebd944e-783d-4fcf-91b4-da5652e69d62 service nova] Lock "85058146-ba4c-4e9a-bbe7-048ed56167ab-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1063.019473] env[63531]: DEBUG nova.compute.manager [req-d029a0d2-c695-4ccb-851a-f090a0ce7dfd req-6ebd944e-783d-4fcf-91b4-da5652e69d62 service nova] [instance: 85058146-ba4c-4e9a-bbe7-048ed56167ab] No waiting events found dispatching network-vif-plugged-9216e469-ea4e-4e8a-9f74-e9310da25110 {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1063.019654] env[63531]: WARNING nova.compute.manager [req-d029a0d2-c695-4ccb-851a-f090a0ce7dfd req-6ebd944e-783d-4fcf-91b4-da5652e69d62 service nova] [instance: 85058146-ba4c-4e9a-bbe7-048ed56167ab] Received unexpected event network-vif-plugged-9216e469-ea4e-4e8a-9f74-e9310da25110 for instance with vm_state building and task_state spawning. [ 1063.067159] env[63531]: DEBUG oslo_concurrency.lockutils [None req-aa491dbe-51b5-432e-ab71-8cdd1877a2f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "97d3f108-e299-4ef1-84ca-532cd64905d5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.380s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1063.143086] env[63531]: DEBUG nova.network.neutron [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 85058146-ba4c-4e9a-bbe7-048ed56167ab] Successfully updated port: 9216e469-ea4e-4e8a-9f74-e9310da25110 {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1063.262076] env[63531]: DEBUG oslo_vmware.api [None req-09413a86-09cb-400a-aa04-8dbb3cff6da3 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118531, 'name': ReconfigVM_Task, 'duration_secs': 8.225699} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.262076] env[63531]: DEBUG oslo_concurrency.lockutils [None req-09413a86-09cb-400a-aa04-8dbb3cff6da3 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Releasing lock "e742d3af-5a45-4274-adfb-c6138bf84d6a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1063.262076] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-09413a86-09cb-400a-aa04-8dbb3cff6da3 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Reconfigured VM to detach interface {{(pid=63531) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 1063.344778] env[63531]: DEBUG nova.virt.hardware [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1063.345064] env[63531]: DEBUG nova.virt.hardware [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1063.345234] env[63531]: DEBUG nova.virt.hardware [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1063.345436] env[63531]: DEBUG nova.virt.hardware [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1063.345587] env[63531]: DEBUG nova.virt.hardware [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1063.345734] env[63531]: DEBUG nova.virt.hardware [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1063.345941] env[63531]: DEBUG nova.virt.hardware [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1063.346120] env[63531]: DEBUG nova.virt.hardware [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1063.346456] env[63531]: DEBUG nova.virt.hardware [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1063.346757] env[63531]: DEBUG nova.virt.hardware [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1063.346865] env[63531]: DEBUG nova.virt.hardware [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1063.356514] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Reconfiguring VM instance instance-0000005c to detach disk 2000 {{(pid=63531) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1063.356896] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-35a52eb1-65fe-4b36-9486-0a7cf540065f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.388154] env[63531]: DEBUG oslo_vmware.api [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118550, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.389951] env[63531]: DEBUG oslo_vmware.api [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 1063.389951] env[63531]: value = "task-1118552" [ 1063.389951] env[63531]: _type = "Task" [ 1063.389951] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.399308] env[63531]: DEBUG oslo_vmware.api [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118552, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.645699] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "refresh_cache-85058146-ba4c-4e9a-bbe7-048ed56167ab" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1063.646050] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquired lock "refresh_cache-85058146-ba4c-4e9a-bbe7-048ed56167ab" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1063.646050] env[63531]: DEBUG nova.network.neutron [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 85058146-ba4c-4e9a-bbe7-048ed56167ab] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1063.726027] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0c5ded2-8f71-447f-be42-b8b3584bb8eb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.733324] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d458716-60c5-4387-b6c1-f7c4009325f2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.762422] env[63531]: DEBUG oslo_concurrency.lockutils [None req-050cb3b4-5bbb-4627-9fc5-b3033fecceb1 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "b6db77e6-4447-4b22-93d7-265ff0ada0bd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.762676] env[63531]: DEBUG oslo_concurrency.lockutils [None req-050cb3b4-5bbb-4627-9fc5-b3033fecceb1 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "b6db77e6-4447-4b22-93d7-265ff0ada0bd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.762884] env[63531]: DEBUG oslo_concurrency.lockutils [None req-050cb3b4-5bbb-4627-9fc5-b3033fecceb1 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "b6db77e6-4447-4b22-93d7-265ff0ada0bd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.763091] env[63531]: DEBUG oslo_concurrency.lockutils [None req-050cb3b4-5bbb-4627-9fc5-b3033fecceb1 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "b6db77e6-4447-4b22-93d7-265ff0ada0bd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.763270] env[63531]: DEBUG oslo_concurrency.lockutils [None req-050cb3b4-5bbb-4627-9fc5-b3033fecceb1 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "b6db77e6-4447-4b22-93d7-265ff0ada0bd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1063.765810] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d7c3850-3bb6-4c97-a38b-21570b8002d9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.768594] env[63531]: INFO nova.compute.manager [None req-050cb3b4-5bbb-4627-9fc5-b3033fecceb1 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b6db77e6-4447-4b22-93d7-265ff0ada0bd] Terminating instance [ 1063.770625] env[63531]: DEBUG nova.compute.manager [None req-050cb3b4-5bbb-4627-9fc5-b3033fecceb1 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b6db77e6-4447-4b22-93d7-265ff0ada0bd] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1063.770823] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-050cb3b4-5bbb-4627-9fc5-b3033fecceb1 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b6db77e6-4447-4b22-93d7-265ff0ada0bd] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1063.771525] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b8d3c34-c34d-4f71-8b91-1834eda7545d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.777218] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4d375c5-b78d-43ad-98b2-633e609b86a9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.782872] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-050cb3b4-5bbb-4627-9fc5-b3033fecceb1 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b6db77e6-4447-4b22-93d7-265ff0ada0bd] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1063.782968] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-555df573-fafb-45d4-854b-02e66937f0f0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.794044] env[63531]: DEBUG nova.compute.provider_tree [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1063.797041] env[63531]: DEBUG oslo_vmware.api [None req-050cb3b4-5bbb-4627-9fc5-b3033fecceb1 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 1063.797041] env[63531]: value = "task-1118553" [ 1063.797041] env[63531]: _type = "Task" [ 1063.797041] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.808397] env[63531]: DEBUG oslo_vmware.api [None req-050cb3b4-5bbb-4627-9fc5-b3033fecceb1 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118553, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.886143] env[63531]: DEBUG oslo_vmware.api [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118550, 'name': PowerOnVM_Task, 'duration_secs': 0.546918} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.886441] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 8f903d83-adbd-4f71-a6f1-037807746248] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1063.886647] env[63531]: INFO nova.compute.manager [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 8f903d83-adbd-4f71-a6f1-037807746248] Took 7.84 seconds to spawn the instance on the hypervisor. [ 1063.886913] env[63531]: DEBUG nova.compute.manager [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 8f903d83-adbd-4f71-a6f1-037807746248] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1063.887721] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c85dcf7-ed31-4b72-9c79-0cc959bad128 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.898739] env[63531]: DEBUG oslo_vmware.api [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118552, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.108808] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-455f3e56-7a45-4cca-bf4f-e08d6600b671 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Volume attach. Driver type: vmdk {{(pid=63531) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1064.109220] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-455f3e56-7a45-4cca-bf4f-e08d6600b671 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244799', 'volume_id': '13d7c941-f906-4020-b545-f2bf1bf8a7a0', 'name': 'volume-13d7c941-f906-4020-b545-f2bf1bf8a7a0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b7b2b7ac-2321-4cf3-b2ed-4877b893af63', 'attached_at': '', 'detached_at': '', 'volume_id': '13d7c941-f906-4020-b545-f2bf1bf8a7a0', 'serial': '13d7c941-f906-4020-b545-f2bf1bf8a7a0'} {{(pid=63531) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1064.110160] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bba9642-c4ff-4e24-b554-f8334a8d9a47 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.128012] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d90f16a0-90d2-426f-9f13-09141c75cda7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.155619] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-455f3e56-7a45-4cca-bf4f-e08d6600b671 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Reconfiguring VM instance instance-0000005d to attach disk [datastore2] volume-13d7c941-f906-4020-b545-f2bf1bf8a7a0/volume-13d7c941-f906-4020-b545-f2bf1bf8a7a0.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1064.157912] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-02a2b7e3-3147-4aa0-b90e-fc06825505fa {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.177699] env[63531]: DEBUG oslo_vmware.api [None req-455f3e56-7a45-4cca-bf4f-e08d6600b671 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Waiting for the task: (returnval){ [ 1064.177699] env[63531]: value = "task-1118554" [ 1064.177699] env[63531]: _type = "Task" [ 1064.177699] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.187218] env[63531]: DEBUG oslo_vmware.api [None req-455f3e56-7a45-4cca-bf4f-e08d6600b671 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118554, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.196346] env[63531]: DEBUG nova.compute.manager [req-d0bf1a32-a62a-419f-971b-5bfde50dcbcb req-40a715c9-b2d6-4802-8928-626125a75fc8 service nova] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Received event network-vif-deleted-4140c945-4976-4856-876d-14507f0a9e17 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1064.196492] env[63531]: INFO nova.compute.manager [req-d0bf1a32-a62a-419f-971b-5bfde50dcbcb req-40a715c9-b2d6-4802-8928-626125a75fc8 service nova] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Neutron deleted interface 4140c945-4976-4856-876d-14507f0a9e17; detaching it from the instance and deleting it from the info cache [ 1064.196704] env[63531]: DEBUG nova.network.neutron [req-d0bf1a32-a62a-419f-971b-5bfde50dcbcb req-40a715c9-b2d6-4802-8928-626125a75fc8 service nova] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Updating instance_info_cache with network_info: [{"id": "a926b26b-b99b-49f2-b7dd-6ca158ead545", "address": "fa:16:3e:99:7c:93", "network": {"id": "caf86428-59a5-4e50-9135-2ab8473af64d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1673994696-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.168", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ba8b284854242c392aec5326e996239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa926b26b-b9", "ovs_interfaceid": "a926b26b-b99b-49f2-b7dd-6ca158ead545", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "5f33c6bf-e31f-422e-8611-c4efb3719681", "address": "fa:16:3e:bc:a4:95", "network": {"id": "caf86428-59a5-4e50-9135-2ab8473af64d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1673994696-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ba8b284854242c392aec5326e996239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5f33c6bf-e3", "ovs_interfaceid": "5f33c6bf-e31f-422e-8611-c4efb3719681", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1064.200612] env[63531]: DEBUG nova.network.neutron [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 85058146-ba4c-4e9a-bbe7-048ed56167ab] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1064.298284] env[63531]: DEBUG nova.scheduler.client.report [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1064.311562] env[63531]: DEBUG oslo_vmware.api [None req-050cb3b4-5bbb-4627-9fc5-b3033fecceb1 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118553, 'name': PowerOffVM_Task, 'duration_secs': 0.305114} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.311855] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-050cb3b4-5bbb-4627-9fc5-b3033fecceb1 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b6db77e6-4447-4b22-93d7-265ff0ada0bd] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1064.312038] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-050cb3b4-5bbb-4627-9fc5-b3033fecceb1 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b6db77e6-4447-4b22-93d7-265ff0ada0bd] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1064.312316] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cd00bf49-f507-45ae-addb-0950f9be494b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.390730] env[63531]: DEBUG nova.network.neutron [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 85058146-ba4c-4e9a-bbe7-048ed56167ab] Updating instance_info_cache with network_info: [{"id": "9216e469-ea4e-4e8a-9f74-e9310da25110", "address": "fa:16:3e:02:6e:66", "network": {"id": "bf99b8cb-9efb-4a74-9db1-0f5f0875f26a", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-998205595-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2e99349303e413b9f7896d449cb7dff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc6e6fe1-c4f5-4389-a49f-0978060eebb4", "external-id": "nsx-vlan-transportzone-829", "segmentation_id": 829, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9216e469-ea", "ovs_interfaceid": "9216e469-ea4e-4e8a-9f74-e9310da25110", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1064.402125] env[63531]: DEBUG oslo_vmware.api [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118552, 'name': ReconfigVM_Task, 'duration_secs': 0.527929} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.403060] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Reconfigured VM instance instance-0000005c to detach disk 2000 {{(pid=63531) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1064.403291] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5dc96b2-1092-4f86-91d4-9852dbfd11c2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.433231] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Reconfiguring VM instance instance-0000005c to attach disk [datastore1] 18057005-1ff1-4e3d-bccc-8c89755d6db7/18057005-1ff1-4e3d-bccc-8c89755d6db7.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1064.433771] env[63531]: INFO nova.compute.manager [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 8f903d83-adbd-4f71-a6f1-037807746248] Took 12.79 seconds to build instance. [ 1064.434826] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0f48bf36-b35d-4261-a184-64cc9291bc1e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.454701] env[63531]: DEBUG oslo_vmware.api [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 1064.454701] env[63531]: value = "task-1118556" [ 1064.454701] env[63531]: _type = "Task" [ 1064.454701] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.463302] env[63531]: DEBUG oslo_vmware.api [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118556, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.537041] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-050cb3b4-5bbb-4627-9fc5-b3033fecceb1 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b6db77e6-4447-4b22-93d7-265ff0ada0bd] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1064.537041] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-050cb3b4-5bbb-4627-9fc5-b3033fecceb1 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b6db77e6-4447-4b22-93d7-265ff0ada0bd] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1064.537041] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-050cb3b4-5bbb-4627-9fc5-b3033fecceb1 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Deleting the datastore file [datastore1] b6db77e6-4447-4b22-93d7-265ff0ada0bd {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1064.537041] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3627e100-5e7c-410d-8840-303cf7526c4a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.544312] env[63531]: DEBUG oslo_vmware.api [None req-050cb3b4-5bbb-4627-9fc5-b3033fecceb1 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 1064.544312] env[63531]: value = "task-1118557" [ 1064.544312] env[63531]: _type = "Task" [ 1064.544312] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.553098] env[63531]: DEBUG oslo_vmware.api [None req-050cb3b4-5bbb-4627-9fc5-b3033fecceb1 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118557, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.657569] env[63531]: DEBUG oslo_concurrency.lockutils [None req-09413a86-09cb-400a-aa04-8dbb3cff6da3 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "refresh_cache-e742d3af-5a45-4274-adfb-c6138bf84d6a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1064.657950] env[63531]: DEBUG oslo_concurrency.lockutils [None req-09413a86-09cb-400a-aa04-8dbb3cff6da3 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquired lock "refresh_cache-e742d3af-5a45-4274-adfb-c6138bf84d6a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.658142] env[63531]: DEBUG nova.network.neutron [None req-09413a86-09cb-400a-aa04-8dbb3cff6da3 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1064.690080] env[63531]: DEBUG oslo_vmware.api [None req-455f3e56-7a45-4cca-bf4f-e08d6600b671 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118554, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.699015] env[63531]: DEBUG oslo_concurrency.lockutils [req-d0bf1a32-a62a-419f-971b-5bfde50dcbcb req-40a715c9-b2d6-4802-8928-626125a75fc8 service nova] Acquiring lock "e742d3af-5a45-4274-adfb-c6138bf84d6a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1064.699262] env[63531]: DEBUG oslo_concurrency.lockutils [req-d0bf1a32-a62a-419f-971b-5bfde50dcbcb req-40a715c9-b2d6-4802-8928-626125a75fc8 service nova] Acquired lock "e742d3af-5a45-4274-adfb-c6138bf84d6a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.700117] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-935d410d-39ab-44f1-87fe-b511acaf3a5f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.720918] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcbe3531-fd9b-4d45-883e-afa62e987e82 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.724260] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c2241b8b-55fa-4beb-a8b2-7758c1bb7c17 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "e742d3af-5a45-4274-adfb-c6138bf84d6a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1064.750459] env[63531]: DEBUG nova.virt.vmwareapi.vmops [req-d0bf1a32-a62a-419f-971b-5bfde50dcbcb req-40a715c9-b2d6-4802-8928-626125a75fc8 service nova] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Reconfiguring VM to detach interface {{(pid=63531) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 1064.750736] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c37ee901-b193-4c5f-8cdf-ae2f11b56ff9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.769774] env[63531]: DEBUG oslo_vmware.api [req-d0bf1a32-a62a-419f-971b-5bfde50dcbcb req-40a715c9-b2d6-4802-8928-626125a75fc8 service nova] Waiting for the task: (returnval){ [ 1064.769774] env[63531]: value = "task-1118558" [ 1064.769774] env[63531]: _type = "Task" [ 1064.769774] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.779260] env[63531]: DEBUG oslo_vmware.api [req-d0bf1a32-a62a-419f-971b-5bfde50dcbcb req-40a715c9-b2d6-4802-8928-626125a75fc8 service nova] Task: {'id': task-1118558, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.807084] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.270s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1064.807084] env[63531]: DEBUG nova.compute.manager [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1064.809767] env[63531]: DEBUG oslo_concurrency.lockutils [None req-584426ce-5cc1-4420-87bd-4a33f7084720 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.805s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1064.809998] env[63531]: DEBUG nova.objects.instance [None req-584426ce-5cc1-4420-87bd-4a33f7084720 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Lazy-loading 'resources' on Instance uuid 3e7b8a01-d930-43cf-9432-0dabc68da20e {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1064.897343] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Releasing lock "refresh_cache-85058146-ba4c-4e9a-bbe7-048ed56167ab" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1064.897623] env[63531]: DEBUG nova.compute.manager [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 85058146-ba4c-4e9a-bbe7-048ed56167ab] Instance network_info: |[{"id": "9216e469-ea4e-4e8a-9f74-e9310da25110", "address": "fa:16:3e:02:6e:66", "network": {"id": "bf99b8cb-9efb-4a74-9db1-0f5f0875f26a", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-998205595-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2e99349303e413b9f7896d449cb7dff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc6e6fe1-c4f5-4389-a49f-0978060eebb4", "external-id": "nsx-vlan-transportzone-829", "segmentation_id": 829, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9216e469-ea", "ovs_interfaceid": "9216e469-ea4e-4e8a-9f74-e9310da25110", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1064.898069] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 85058146-ba4c-4e9a-bbe7-048ed56167ab] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:02:6e:66', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bc6e6fe1-c4f5-4389-a49f-0978060eebb4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9216e469-ea4e-4e8a-9f74-e9310da25110', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1064.905511] env[63531]: DEBUG oslo.service.loopingcall [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1064.905724] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 85058146-ba4c-4e9a-bbe7-048ed56167ab] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1064.905945] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-beac2ac9-5a42-4534-80ca-9818d3d00316 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.933414] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1064.933414] env[63531]: value = "task-1118559" [ 1064.933414] env[63531]: _type = "Task" [ 1064.933414] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.942317] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118559, 'name': CreateVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.948993] env[63531]: DEBUG oslo_concurrency.lockutils [None req-62ab3959-9ec7-4163-9275-0b31d0e4b825 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "8f903d83-adbd-4f71-a6f1-037807746248" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.311s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1064.965190] env[63531]: DEBUG oslo_vmware.api [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118556, 'name': ReconfigVM_Task, 'duration_secs': 0.406291} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.965658] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Reconfigured VM instance instance-0000005c to attach disk [datastore1] 18057005-1ff1-4e3d-bccc-8c89755d6db7/18057005-1ff1-4e3d-bccc-8c89755d6db7.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1064.966097] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Updating instance '18057005-1ff1-4e3d-bccc-8c89755d6db7' progress to 50 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1065.055465] env[63531]: DEBUG oslo_vmware.api [None req-050cb3b4-5bbb-4627-9fc5-b3033fecceb1 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118557, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.179164} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.055747] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-050cb3b4-5bbb-4627-9fc5-b3033fecceb1 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1065.056197] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-050cb3b4-5bbb-4627-9fc5-b3033fecceb1 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b6db77e6-4447-4b22-93d7-265ff0ada0bd] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1065.056616] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-050cb3b4-5bbb-4627-9fc5-b3033fecceb1 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b6db77e6-4447-4b22-93d7-265ff0ada0bd] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1065.056895] env[63531]: INFO nova.compute.manager [None req-050cb3b4-5bbb-4627-9fc5-b3033fecceb1 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b6db77e6-4447-4b22-93d7-265ff0ada0bd] Took 1.29 seconds to destroy the instance on the hypervisor. [ 1065.057199] env[63531]: DEBUG oslo.service.loopingcall [None req-050cb3b4-5bbb-4627-9fc5-b3033fecceb1 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1065.057423] env[63531]: DEBUG nova.compute.manager [-] [instance: b6db77e6-4447-4b22-93d7-265ff0ada0bd] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1065.057681] env[63531]: DEBUG nova.network.neutron [-] [instance: b6db77e6-4447-4b22-93d7-265ff0ada0bd] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1065.081266] env[63531]: DEBUG nova.compute.manager [req-12df086a-aaee-4fc0-9c79-a53119d48317 req-8bea5b09-c55f-421c-be42-47aed0180cfb service nova] [instance: 85058146-ba4c-4e9a-bbe7-048ed56167ab] Received event network-changed-9216e469-ea4e-4e8a-9f74-e9310da25110 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1065.081471] env[63531]: DEBUG nova.compute.manager [req-12df086a-aaee-4fc0-9c79-a53119d48317 req-8bea5b09-c55f-421c-be42-47aed0180cfb service nova] [instance: 85058146-ba4c-4e9a-bbe7-048ed56167ab] Refreshing instance network info cache due to event network-changed-9216e469-ea4e-4e8a-9f74-e9310da25110. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1065.081727] env[63531]: DEBUG oslo_concurrency.lockutils [req-12df086a-aaee-4fc0-9c79-a53119d48317 req-8bea5b09-c55f-421c-be42-47aed0180cfb service nova] Acquiring lock "refresh_cache-85058146-ba4c-4e9a-bbe7-048ed56167ab" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1065.081892] env[63531]: DEBUG oslo_concurrency.lockutils [req-12df086a-aaee-4fc0-9c79-a53119d48317 req-8bea5b09-c55f-421c-be42-47aed0180cfb service nova] Acquired lock "refresh_cache-85058146-ba4c-4e9a-bbe7-048ed56167ab" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1065.082105] env[63531]: DEBUG nova.network.neutron [req-12df086a-aaee-4fc0-9c79-a53119d48317 req-8bea5b09-c55f-421c-be42-47aed0180cfb service nova] [instance: 85058146-ba4c-4e9a-bbe7-048ed56167ab] Refreshing network info cache for port 9216e469-ea4e-4e8a-9f74-e9310da25110 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1065.194974] env[63531]: DEBUG oslo_vmware.api [None req-455f3e56-7a45-4cca-bf4f-e08d6600b671 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118554, 'name': ReconfigVM_Task, 'duration_secs': 0.572608} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.195334] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-455f3e56-7a45-4cca-bf4f-e08d6600b671 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Reconfigured VM instance instance-0000005d to attach disk [datastore2] volume-13d7c941-f906-4020-b545-f2bf1bf8a7a0/volume-13d7c941-f906-4020-b545-f2bf1bf8a7a0.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1065.202700] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5fc4ef4e-253f-42d4-8ebd-0bbc9eb8dd31 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.226113] env[63531]: DEBUG oslo_vmware.api [None req-455f3e56-7a45-4cca-bf4f-e08d6600b671 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Waiting for the task: (returnval){ [ 1065.226113] env[63531]: value = "task-1118560" [ 1065.226113] env[63531]: _type = "Task" [ 1065.226113] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.241047] env[63531]: DEBUG oslo_vmware.api [None req-455f3e56-7a45-4cca-bf4f-e08d6600b671 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118560, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.281203] env[63531]: DEBUG oslo_vmware.api [req-d0bf1a32-a62a-419f-971b-5bfde50dcbcb req-40a715c9-b2d6-4802-8928-626125a75fc8 service nova] Task: {'id': task-1118558, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.312887] env[63531]: DEBUG nova.compute.utils [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1065.317349] env[63531]: DEBUG nova.compute.manager [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1065.317525] env[63531]: DEBUG nova.network.neutron [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1065.367141] env[63531]: DEBUG nova.policy [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1ea586d8e4f947d2902c130530d42466', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f9056486e3b640d0bcdb9369ed5d67ad', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 1065.446137] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118559, 'name': CreateVM_Task, 'duration_secs': 0.47409} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.446320] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 85058146-ba4c-4e9a-bbe7-048ed56167ab] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1065.447032] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1065.447220] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1065.447541] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1065.447820] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-75700f75-5255-40a7-97be-5d15575a9109 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.453911] env[63531]: DEBUG oslo_vmware.api [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 1065.453911] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52e56360-d0f6-6057-707a-9c49fa2115bd" [ 1065.453911] env[63531]: _type = "Task" [ 1065.453911] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.465193] env[63531]: DEBUG oslo_vmware.api [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52e56360-d0f6-6057-707a-9c49fa2115bd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.474150] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d50b6ebb-191f-4d3f-8972-0f02d23b52c2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.497977] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0e84ce7-4539-4cae-b061-866e3a7cac06 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.520493] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Updating instance '18057005-1ff1-4e3d-bccc-8c89755d6db7' progress to 67 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1065.577830] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3451c82-1489-4af2-b7b1-cb4e3cd44669 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.589402] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7237eba3-e885-4d8f-b0e4-43bdc437499b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.622987] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c8c9c26-1be3-49f3-b7f6-33e9a5fac2cf {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.631737] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2288412b-17b3-4647-824c-2bb794810845 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.646019] env[63531]: DEBUG nova.compute.provider_tree [None req-584426ce-5cc1-4420-87bd-4a33f7084720 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1065.720522] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f98a7638-6ab3-43f3-8fc7-8daec4306176 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "8f903d83-adbd-4f71-a6f1-037807746248" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1065.720767] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f98a7638-6ab3-43f3-8fc7-8daec4306176 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "8f903d83-adbd-4f71-a6f1-037807746248" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1065.720951] env[63531]: DEBUG nova.compute.manager [None req-f98a7638-6ab3-43f3-8fc7-8daec4306176 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 8f903d83-adbd-4f71-a6f1-037807746248] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1065.722192] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d24664de-03d7-43c6-95fb-35d645f75a50 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.731542] env[63531]: DEBUG nova.compute.manager [None req-f98a7638-6ab3-43f3-8fc7-8daec4306176 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 8f903d83-adbd-4f71-a6f1-037807746248] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63531) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1065.732626] env[63531]: DEBUG nova.objects.instance [None req-f98a7638-6ab3-43f3-8fc7-8daec4306176 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lazy-loading 'flavor' on Instance uuid 8f903d83-adbd-4f71-a6f1-037807746248 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1065.739341] env[63531]: DEBUG oslo_vmware.api [None req-455f3e56-7a45-4cca-bf4f-e08d6600b671 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118560, 'name': ReconfigVM_Task, 'duration_secs': 0.163008} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.739341] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-455f3e56-7a45-4cca-bf4f-e08d6600b671 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244799', 'volume_id': '13d7c941-f906-4020-b545-f2bf1bf8a7a0', 'name': 'volume-13d7c941-f906-4020-b545-f2bf1bf8a7a0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b7b2b7ac-2321-4cf3-b2ed-4877b893af63', 'attached_at': '', 'detached_at': '', 'volume_id': '13d7c941-f906-4020-b545-f2bf1bf8a7a0', 'serial': '13d7c941-f906-4020-b545-f2bf1bf8a7a0'} {{(pid=63531) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1065.772667] env[63531]: DEBUG nova.network.neutron [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Successfully created port: bb17faa5-c291-4bfa-b2fa-c435a815dc3d {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1065.789302] env[63531]: DEBUG oslo_vmware.api [req-d0bf1a32-a62a-419f-971b-5bfde50dcbcb req-40a715c9-b2d6-4802-8928-626125a75fc8 service nova] Task: {'id': task-1118558, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.817610] env[63531]: DEBUG nova.compute.manager [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1065.948212] env[63531]: INFO nova.network.neutron [None req-09413a86-09cb-400a-aa04-8dbb3cff6da3 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Port 5f33c6bf-e31f-422e-8611-c4efb3719681 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1065.948738] env[63531]: DEBUG nova.network.neutron [None req-09413a86-09cb-400a-aa04-8dbb3cff6da3 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Updating instance_info_cache with network_info: [{"id": "a926b26b-b99b-49f2-b7dd-6ca158ead545", "address": "fa:16:3e:99:7c:93", "network": {"id": "caf86428-59a5-4e50-9135-2ab8473af64d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1673994696-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.168", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ba8b284854242c392aec5326e996239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa926b26b-b9", "ovs_interfaceid": "a926b26b-b99b-49f2-b7dd-6ca158ead545", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1065.956643] env[63531]: DEBUG nova.network.neutron [req-12df086a-aaee-4fc0-9c79-a53119d48317 req-8bea5b09-c55f-421c-be42-47aed0180cfb service nova] [instance: 85058146-ba4c-4e9a-bbe7-048ed56167ab] Updated VIF entry in instance network info cache for port 9216e469-ea4e-4e8a-9f74-e9310da25110. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1065.957107] env[63531]: DEBUG nova.network.neutron [req-12df086a-aaee-4fc0-9c79-a53119d48317 req-8bea5b09-c55f-421c-be42-47aed0180cfb service nova] [instance: 85058146-ba4c-4e9a-bbe7-048ed56167ab] Updating instance_info_cache with network_info: [{"id": "9216e469-ea4e-4e8a-9f74-e9310da25110", "address": "fa:16:3e:02:6e:66", "network": {"id": "bf99b8cb-9efb-4a74-9db1-0f5f0875f26a", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-998205595-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2e99349303e413b9f7896d449cb7dff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc6e6fe1-c4f5-4389-a49f-0978060eebb4", "external-id": "nsx-vlan-transportzone-829", "segmentation_id": 829, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9216e469-ea", "ovs_interfaceid": "9216e469-ea4e-4e8a-9f74-e9310da25110", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1065.974558] env[63531]: DEBUG oslo_vmware.api [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52e56360-d0f6-6057-707a-9c49fa2115bd, 'name': SearchDatastore_Task, 'duration_secs': 0.011417} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.975119] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1065.975568] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 85058146-ba4c-4e9a-bbe7-048ed56167ab] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1065.976033] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1065.976313] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1065.976617] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1065.977259] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0291d23d-0e1c-406e-9dba-cf3099bcc85a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.989034] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1065.989034] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1065.989504] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0a9a4a5b-d41a-4ba9-bef7-820d20561e8f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.996371] env[63531]: DEBUG oslo_vmware.api [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 1065.996371] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]525ecce1-2129-bb5d-69a0-df1c1f720c77" [ 1065.996371] env[63531]: _type = "Task" [ 1065.996371] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.005844] env[63531]: DEBUG oslo_vmware.api [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]525ecce1-2129-bb5d-69a0-df1c1f720c77, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.149388] env[63531]: DEBUG nova.scheduler.client.report [None req-584426ce-5cc1-4420-87bd-4a33f7084720 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1066.237286] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-f98a7638-6ab3-43f3-8fc7-8daec4306176 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 8f903d83-adbd-4f71-a6f1-037807746248] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1066.237665] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-df6d488f-494b-49c1-9981-9aa9c671897d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.246825] env[63531]: DEBUG oslo_vmware.api [None req-f98a7638-6ab3-43f3-8fc7-8daec4306176 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 1066.246825] env[63531]: value = "task-1118562" [ 1066.246825] env[63531]: _type = "Task" [ 1066.246825] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.255574] env[63531]: DEBUG oslo_vmware.api [None req-f98a7638-6ab3-43f3-8fc7-8daec4306176 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118562, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.278693] env[63531]: DEBUG nova.network.neutron [-] [instance: b6db77e6-4447-4b22-93d7-265ff0ada0bd] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1066.286566] env[63531]: DEBUG oslo_vmware.api [req-d0bf1a32-a62a-419f-971b-5bfde50dcbcb req-40a715c9-b2d6-4802-8928-626125a75fc8 service nova] Task: {'id': task-1118558, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.456681] env[63531]: DEBUG oslo_concurrency.lockutils [None req-09413a86-09cb-400a-aa04-8dbb3cff6da3 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Releasing lock "refresh_cache-e742d3af-5a45-4274-adfb-c6138bf84d6a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1066.463991] env[63531]: DEBUG oslo_concurrency.lockutils [req-12df086a-aaee-4fc0-9c79-a53119d48317 req-8bea5b09-c55f-421c-be42-47aed0180cfb service nova] Releasing lock "refresh_cache-85058146-ba4c-4e9a-bbe7-048ed56167ab" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1066.464381] env[63531]: DEBUG nova.compute.manager [req-12df086a-aaee-4fc0-9c79-a53119d48317 req-8bea5b09-c55f-421c-be42-47aed0180cfb service nova] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Received event network-vif-deleted-5f33c6bf-e31f-422e-8611-c4efb3719681 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1066.464515] env[63531]: INFO nova.compute.manager [req-12df086a-aaee-4fc0-9c79-a53119d48317 req-8bea5b09-c55f-421c-be42-47aed0180cfb service nova] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Neutron deleted interface 5f33c6bf-e31f-422e-8611-c4efb3719681; detaching it from the instance and deleting it from the info cache [ 1066.465062] env[63531]: DEBUG nova.network.neutron [req-12df086a-aaee-4fc0-9c79-a53119d48317 req-8bea5b09-c55f-421c-be42-47aed0180cfb service nova] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Updating instance_info_cache with network_info: [{"id": "a926b26b-b99b-49f2-b7dd-6ca158ead545", "address": "fa:16:3e:99:7c:93", "network": {"id": "caf86428-59a5-4e50-9135-2ab8473af64d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1673994696-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.168", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ba8b284854242c392aec5326e996239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa926b26b-b9", "ovs_interfaceid": "a926b26b-b99b-49f2-b7dd-6ca158ead545", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1066.507704] env[63531]: DEBUG oslo_vmware.api [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]525ecce1-2129-bb5d-69a0-df1c1f720c77, 'name': SearchDatastore_Task, 'duration_secs': 0.010287} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.508516] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-27aec8bc-e981-4c14-a7cc-a34e33e7d354 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.514112] env[63531]: DEBUG oslo_vmware.api [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 1066.514112] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52031a18-46fd-4626-b628-f3bc2256ac51" [ 1066.514112] env[63531]: _type = "Task" [ 1066.514112] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.521629] env[63531]: DEBUG oslo_vmware.api [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52031a18-46fd-4626-b628-f3bc2256ac51, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.656373] env[63531]: DEBUG oslo_concurrency.lockutils [None req-584426ce-5cc1-4420-87bd-4a33f7084720 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.846s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.659088] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5dff6cb1-9c4d-4e3f-bec7-275c4d5265b6 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.386s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1066.659370] env[63531]: DEBUG nova.objects.instance [None req-5dff6cb1-9c4d-4e3f-bec7-275c4d5265b6 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Lazy-loading 'resources' on Instance uuid 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1066.680145] env[63531]: INFO nova.scheduler.client.report [None req-584426ce-5cc1-4420-87bd-4a33f7084720 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Deleted allocations for instance 3e7b8a01-d930-43cf-9432-0dabc68da20e [ 1066.756942] env[63531]: DEBUG oslo_vmware.api [None req-f98a7638-6ab3-43f3-8fc7-8daec4306176 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118562, 'name': PowerOffVM_Task} progress is 100%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.776286] env[63531]: DEBUG nova.objects.instance [None req-455f3e56-7a45-4cca-bf4f-e08d6600b671 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lazy-loading 'flavor' on Instance uuid b7b2b7ac-2321-4cf3-b2ed-4877b893af63 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1066.781038] env[63531]: INFO nova.compute.manager [-] [instance: b6db77e6-4447-4b22-93d7-265ff0ada0bd] Took 1.72 seconds to deallocate network for instance. [ 1066.789173] env[63531]: DEBUG oslo_vmware.api [req-d0bf1a32-a62a-419f-971b-5bfde50dcbcb req-40a715c9-b2d6-4802-8928-626125a75fc8 service nova] Task: {'id': task-1118558, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.828983] env[63531]: DEBUG nova.compute.manager [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1066.853832] env[63531]: DEBUG nova.virt.hardware [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1066.854100] env[63531]: DEBUG nova.virt.hardware [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1066.854269] env[63531]: DEBUG nova.virt.hardware [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1066.854468] env[63531]: DEBUG nova.virt.hardware [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1066.854721] env[63531]: DEBUG nova.virt.hardware [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1066.854905] env[63531]: DEBUG nova.virt.hardware [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1066.855136] env[63531]: DEBUG nova.virt.hardware [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1066.855309] env[63531]: DEBUG nova.virt.hardware [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1066.855481] env[63531]: DEBUG nova.virt.hardware [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1066.855655] env[63531]: DEBUG nova.virt.hardware [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1066.855833] env[63531]: DEBUG nova.virt.hardware [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1066.856675] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81e0826c-4a8a-431f-bb3d-42eb4db331a8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.864922] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb9b2ff7-ffb0-4e65-85d0-1df247ac2e09 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.960998] env[63531]: DEBUG oslo_concurrency.lockutils [None req-09413a86-09cb-400a-aa04-8dbb3cff6da3 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "interface-e742d3af-5a45-4274-adfb-c6138bf84d6a-5f33c6bf-e31f-422e-8611-c4efb3719681" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 12.663s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.968223] env[63531]: DEBUG oslo_concurrency.lockutils [req-12df086a-aaee-4fc0-9c79-a53119d48317 req-8bea5b09-c55f-421c-be42-47aed0180cfb service nova] Acquiring lock "e742d3af-5a45-4274-adfb-c6138bf84d6a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1067.024708] env[63531]: DEBUG oslo_vmware.api [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52031a18-46fd-4626-b628-f3bc2256ac51, 'name': SearchDatastore_Task, 'duration_secs': 0.009736} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.024961] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1067.025247] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] 85058146-ba4c-4e9a-bbe7-048ed56167ab/85058146-ba4c-4e9a-bbe7-048ed56167ab.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1067.025502] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3c0beb1d-9425-4f2e-a63b-0731fda93a88 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.032139] env[63531]: DEBUG oslo_vmware.api [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 1067.032139] env[63531]: value = "task-1118563" [ 1067.032139] env[63531]: _type = "Task" [ 1067.032139] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.040942] env[63531]: DEBUG oslo_vmware.api [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118563, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.084116] env[63531]: DEBUG oslo_concurrency.lockutils [None req-71b24bc9-52e0-4981-9ef8-548fbd317f37 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Acquiring lock "b7b2b7ac-2321-4cf3-b2ed-4877b893af63" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.151368] env[63531]: DEBUG nova.compute.manager [req-df143532-23c3-49f4-86cd-8639a137cf00 req-3ea36036-1fc6-496c-9c46-adf865e8dba2 service nova] [instance: b6db77e6-4447-4b22-93d7-265ff0ada0bd] Received event network-vif-deleted-d6c4d45f-8dc2-4f95-947f-9320df584daf {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1067.192442] env[63531]: DEBUG oslo_concurrency.lockutils [None req-584426ce-5cc1-4420-87bd-4a33f7084720 tempest-ServersAaction247Test-339014916 tempest-ServersAaction247Test-339014916-project-member] Lock "3e7b8a01-d930-43cf-9432-0dabc68da20e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.005s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.205286] env[63531]: DEBUG nova.network.neutron [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Port af2485f6-5ce2-4a83-9d95-7bea85cebff2 binding to destination host cpu-1 is already ACTIVE {{(pid=63531) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1067.262155] env[63531]: DEBUG oslo_vmware.api [None req-f98a7638-6ab3-43f3-8fc7-8daec4306176 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118562, 'name': PowerOffVM_Task} progress is 100%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.286167] env[63531]: DEBUG oslo_concurrency.lockutils [None req-455f3e56-7a45-4cca-bf4f-e08d6600b671 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lock "b7b2b7ac-2321-4cf3-b2ed-4877b893af63" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.786s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.287045] env[63531]: DEBUG oslo_concurrency.lockutils [None req-71b24bc9-52e0-4981-9ef8-548fbd317f37 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lock "b7b2b7ac-2321-4cf3-b2ed-4877b893af63" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.203s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.290808] env[63531]: DEBUG oslo_concurrency.lockutils [None req-050cb3b4-5bbb-4627-9fc5-b3033fecceb1 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.299992] env[63531]: DEBUG oslo_vmware.api [req-d0bf1a32-a62a-419f-971b-5bfde50dcbcb req-40a715c9-b2d6-4802-8928-626125a75fc8 service nova] Task: {'id': task-1118558, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.390280] env[63531]: DEBUG nova.compute.manager [req-d9afca27-f37f-449f-bacb-e291bb2aefd8 req-de3d2bf9-8230-4ece-873c-9d3fcdfb46d5 service nova] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Received event network-vif-plugged-bb17faa5-c291-4bfa-b2fa-c435a815dc3d {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1067.390535] env[63531]: DEBUG oslo_concurrency.lockutils [req-d9afca27-f37f-449f-bacb-e291bb2aefd8 req-de3d2bf9-8230-4ece-873c-9d3fcdfb46d5 service nova] Acquiring lock "89f19669-53a9-4425-a3c6-d5f8873b707e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.390797] env[63531]: DEBUG oslo_concurrency.lockutils [req-d9afca27-f37f-449f-bacb-e291bb2aefd8 req-de3d2bf9-8230-4ece-873c-9d3fcdfb46d5 service nova] Lock "89f19669-53a9-4425-a3c6-d5f8873b707e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.390936] env[63531]: DEBUG oslo_concurrency.lockutils [req-d9afca27-f37f-449f-bacb-e291bb2aefd8 req-de3d2bf9-8230-4ece-873c-9d3fcdfb46d5 service nova] Lock "89f19669-53a9-4425-a3c6-d5f8873b707e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.391957] env[63531]: DEBUG nova.compute.manager [req-d9afca27-f37f-449f-bacb-e291bb2aefd8 req-de3d2bf9-8230-4ece-873c-9d3fcdfb46d5 service nova] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] No waiting events found dispatching network-vif-plugged-bb17faa5-c291-4bfa-b2fa-c435a815dc3d {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1067.392513] env[63531]: WARNING nova.compute.manager [req-d9afca27-f37f-449f-bacb-e291bb2aefd8 req-de3d2bf9-8230-4ece-873c-9d3fcdfb46d5 service nova] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Received unexpected event network-vif-plugged-bb17faa5-c291-4bfa-b2fa-c435a815dc3d for instance with vm_state building and task_state spawning. [ 1067.450304] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a9c5ea5-4cfd-4317-9484-b8512c62da4f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.458369] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b424257-7107-4146-8727-e59fd3214a3d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.495818] env[63531]: DEBUG nova.network.neutron [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Successfully updated port: bb17faa5-c291-4bfa-b2fa-c435a815dc3d {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1067.498269] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92726766-6c5c-42f2-a4f6-0eb5d606580b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.508736] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d260bb88-6f33-4741-8bec-80d5696210ff {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.526395] env[63531]: DEBUG nova.compute.provider_tree [None req-5dff6cb1-9c4d-4e3f-bec7-275c4d5265b6 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1067.543800] env[63531]: DEBUG oslo_vmware.api [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118563, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.473828} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.544090] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] 85058146-ba4c-4e9a-bbe7-048ed56167ab/85058146-ba4c-4e9a-bbe7-048ed56167ab.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1067.544316] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 85058146-ba4c-4e9a-bbe7-048ed56167ab] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1067.544567] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cf8f2c9f-6684-40e9-abfa-a510366aaa56 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.553828] env[63531]: DEBUG oslo_vmware.api [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 1067.553828] env[63531]: value = "task-1118565" [ 1067.553828] env[63531]: _type = "Task" [ 1067.553828] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.565075] env[63531]: DEBUG oslo_vmware.api [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118565, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.761191] env[63531]: DEBUG oslo_vmware.api [None req-f98a7638-6ab3-43f3-8fc7-8daec4306176 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118562, 'name': PowerOffVM_Task, 'duration_secs': 1.181233} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.762378] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-f98a7638-6ab3-43f3-8fc7-8daec4306176 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 8f903d83-adbd-4f71-a6f1-037807746248] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1067.762571] env[63531]: DEBUG nova.compute.manager [None req-f98a7638-6ab3-43f3-8fc7-8daec4306176 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 8f903d83-adbd-4f71-a6f1-037807746248] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1067.764172] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-506c05ca-2a99-4a1e-9c74-90d9c3a49f60 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.789072] env[63531]: DEBUG oslo_vmware.api [req-d0bf1a32-a62a-419f-971b-5bfde50dcbcb req-40a715c9-b2d6-4802-8928-626125a75fc8 service nova] Task: {'id': task-1118558, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.797323] env[63531]: INFO nova.compute.manager [None req-71b24bc9-52e0-4981-9ef8-548fbd317f37 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Detaching volume a34e5b41-b9fc-4c5a-967e-b3e601dde931 [ 1067.829128] env[63531]: INFO nova.virt.block_device [None req-71b24bc9-52e0-4981-9ef8-548fbd317f37 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Attempting to driver detach volume a34e5b41-b9fc-4c5a-967e-b3e601dde931 from mountpoint /dev/sdb [ 1067.829480] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-71b24bc9-52e0-4981-9ef8-548fbd317f37 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Volume detach. Driver type: vmdk {{(pid=63531) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1067.829707] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-71b24bc9-52e0-4981-9ef8-548fbd317f37 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244796', 'volume_id': 'a34e5b41-b9fc-4c5a-967e-b3e601dde931', 'name': 'volume-a34e5b41-b9fc-4c5a-967e-b3e601dde931', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b7b2b7ac-2321-4cf3-b2ed-4877b893af63', 'attached_at': '', 'detached_at': '', 'volume_id': 'a34e5b41-b9fc-4c5a-967e-b3e601dde931', 'serial': 'a34e5b41-b9fc-4c5a-967e-b3e601dde931'} {{(pid=63531) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1067.830640] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aafc443f-aa5e-4a82-97a7-004791ae17d8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.857056] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e41d52b4-7888-4cc1-ab21-9b6e419afc41 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.865263] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0290271f-8f6c-4483-9eec-98ab09bae2e0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.893138] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bbaf852-4589-4bea-9d85-d9f829685fd0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.909768] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-71b24bc9-52e0-4981-9ef8-548fbd317f37 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] The volume has not been displaced from its original location: [datastore2] volume-a34e5b41-b9fc-4c5a-967e-b3e601dde931/volume-a34e5b41-b9fc-4c5a-967e-b3e601dde931.vmdk. No consolidation needed. {{(pid=63531) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1067.915155] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-71b24bc9-52e0-4981-9ef8-548fbd317f37 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Reconfiguring VM instance instance-0000005d to detach disk 2001 {{(pid=63531) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1067.915922] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7e18ee40-21fd-416a-98fe-c59350a5ecef {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.939222] env[63531]: DEBUG oslo_vmware.api [None req-71b24bc9-52e0-4981-9ef8-548fbd317f37 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Waiting for the task: (returnval){ [ 1067.939222] env[63531]: value = "task-1118566" [ 1067.939222] env[63531]: _type = "Task" [ 1067.939222] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.947515] env[63531]: DEBUG oslo_vmware.api [None req-71b24bc9-52e0-4981-9ef8-548fbd317f37 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118566, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.003522] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquiring lock "refresh_cache-89f19669-53a9-4425-a3c6-d5f8873b707e" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1068.003748] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquired lock "refresh_cache-89f19669-53a9-4425-a3c6-d5f8873b707e" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1068.003978] env[63531]: DEBUG nova.network.neutron [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1068.030134] env[63531]: DEBUG nova.scheduler.client.report [None req-5dff6cb1-9c4d-4e3f-bec7-275c4d5265b6 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1068.064625] env[63531]: DEBUG oslo_vmware.api [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118565, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.058436} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.066110] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 85058146-ba4c-4e9a-bbe7-048ed56167ab] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1068.069735] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29b46b41-a9ea-475b-8c7f-b9596f66b055 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.094756] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 85058146-ba4c-4e9a-bbe7-048ed56167ab] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] 85058146-ba4c-4e9a-bbe7-048ed56167ab/85058146-ba4c-4e9a-bbe7-048ed56167ab.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1068.096052] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-02bd5a30-80d6-482f-882b-15548d0a8d22 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.117605] env[63531]: DEBUG oslo_vmware.api [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 1068.117605] env[63531]: value = "task-1118567" [ 1068.117605] env[63531]: _type = "Task" [ 1068.117605] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.127725] env[63531]: DEBUG oslo_vmware.api [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118567, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.236435] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquiring lock "18057005-1ff1-4e3d-bccc-8c89755d6db7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.236703] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "18057005-1ff1-4e3d-bccc-8c89755d6db7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.236872] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "18057005-1ff1-4e3d-bccc-8c89755d6db7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.278108] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f98a7638-6ab3-43f3-8fc7-8daec4306176 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "8f903d83-adbd-4f71-a6f1-037807746248" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.557s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.288753] env[63531]: DEBUG oslo_vmware.api [req-d0bf1a32-a62a-419f-971b-5bfde50dcbcb req-40a715c9-b2d6-4802-8928-626125a75fc8 service nova] Task: {'id': task-1118558, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.451962] env[63531]: DEBUG oslo_vmware.api [None req-71b24bc9-52e0-4981-9ef8-548fbd317f37 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118566, 'name': ReconfigVM_Task, 'duration_secs': 0.259725} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.452357] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-71b24bc9-52e0-4981-9ef8-548fbd317f37 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Reconfigured VM instance instance-0000005d to detach disk 2001 {{(pid=63531) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1068.457431] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9585c287-0003-484f-8f34-38f1961522f5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.475733] env[63531]: DEBUG oslo_vmware.api [None req-71b24bc9-52e0-4981-9ef8-548fbd317f37 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Waiting for the task: (returnval){ [ 1068.475733] env[63531]: value = "task-1118568" [ 1068.475733] env[63531]: _type = "Task" [ 1068.475733] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.485432] env[63531]: DEBUG oslo_vmware.api [None req-71b24bc9-52e0-4981-9ef8-548fbd317f37 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118568, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.535970] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5dff6cb1-9c4d-4e3f-bec7-275c4d5265b6 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.877s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.538454] env[63531]: DEBUG oslo_concurrency.lockutils [None req-050cb3b4-5bbb-4627-9fc5-b3033fecceb1 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.248s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.538698] env[63531]: DEBUG nova.objects.instance [None req-050cb3b4-5bbb-4627-9fc5-b3033fecceb1 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lazy-loading 'resources' on Instance uuid b6db77e6-4447-4b22-93d7-265ff0ada0bd {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1068.547525] env[63531]: DEBUG nova.network.neutron [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1068.560028] env[63531]: INFO nova.scheduler.client.report [None req-5dff6cb1-9c4d-4e3f-bec7-275c4d5265b6 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Deleted allocations for instance 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df [ 1068.574016] env[63531]: DEBUG oslo_concurrency.lockutils [None req-19dcebb4-f9d8-49dd-bf4f-f27015ebfdab tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "8f903d83-adbd-4f71-a6f1-037807746248" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.574313] env[63531]: DEBUG oslo_concurrency.lockutils [None req-19dcebb4-f9d8-49dd-bf4f-f27015ebfdab tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "8f903d83-adbd-4f71-a6f1-037807746248" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.574541] env[63531]: DEBUG oslo_concurrency.lockutils [None req-19dcebb4-f9d8-49dd-bf4f-f27015ebfdab tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "8f903d83-adbd-4f71-a6f1-037807746248-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.574741] env[63531]: DEBUG oslo_concurrency.lockutils [None req-19dcebb4-f9d8-49dd-bf4f-f27015ebfdab tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "8f903d83-adbd-4f71-a6f1-037807746248-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.574904] env[63531]: DEBUG oslo_concurrency.lockutils [None req-19dcebb4-f9d8-49dd-bf4f-f27015ebfdab tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "8f903d83-adbd-4f71-a6f1-037807746248-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.577175] env[63531]: INFO nova.compute.manager [None req-19dcebb4-f9d8-49dd-bf4f-f27015ebfdab tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 8f903d83-adbd-4f71-a6f1-037807746248] Terminating instance [ 1068.579105] env[63531]: DEBUG nova.compute.manager [None req-19dcebb4-f9d8-49dd-bf4f-f27015ebfdab tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 8f903d83-adbd-4f71-a6f1-037807746248] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1068.579319] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-19dcebb4-f9d8-49dd-bf4f-f27015ebfdab tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 8f903d83-adbd-4f71-a6f1-037807746248] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1068.580341] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d06340ed-be38-4796-95c5-fc723ed7df54 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.589453] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-19dcebb4-f9d8-49dd-bf4f-f27015ebfdab tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 8f903d83-adbd-4f71-a6f1-037807746248] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1068.589739] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b7a8cd8b-786e-48cd-bfba-9c069e8302dd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.630073] env[63531]: DEBUG oslo_vmware.api [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118567, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.671046] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-19dcebb4-f9d8-49dd-bf4f-f27015ebfdab tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 8f903d83-adbd-4f71-a6f1-037807746248] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1068.671280] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-19dcebb4-f9d8-49dd-bf4f-f27015ebfdab tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 8f903d83-adbd-4f71-a6f1-037807746248] Deleting contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1068.671553] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-19dcebb4-f9d8-49dd-bf4f-f27015ebfdab tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Deleting the datastore file [datastore2] 8f903d83-adbd-4f71-a6f1-037807746248 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1068.672194] env[63531]: DEBUG nova.compute.manager [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Stashing vm_state: active {{(pid=63531) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1068.675491] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-719a17ef-51c4-47a0-a40e-4c92bb754f0d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.684728] env[63531]: DEBUG oslo_vmware.api [None req-19dcebb4-f9d8-49dd-bf4f-f27015ebfdab tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 1068.684728] env[63531]: value = "task-1118570" [ 1068.684728] env[63531]: _type = "Task" [ 1068.684728] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.694471] env[63531]: DEBUG oslo_vmware.api [None req-19dcebb4-f9d8-49dd-bf4f-f27015ebfdab tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118570, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.720100] env[63531]: DEBUG nova.network.neutron [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Updating instance_info_cache with network_info: [{"id": "bb17faa5-c291-4bfa-b2fa-c435a815dc3d", "address": "fa:16:3e:2e:a7:41", "network": {"id": "95ed2c7d-3d1a-4736-8e90-104002f5c40b", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-73549344-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9056486e3b640d0bcdb9369ed5d67ad", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04e15990-16e1-4cb2-b0f0-06c362e68c5e", "external-id": "nsx-vlan-transportzone-555", "segmentation_id": 555, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb17faa5-c2", "ovs_interfaceid": "bb17faa5-c291-4bfa-b2fa-c435a815dc3d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1068.790728] env[63531]: DEBUG oslo_vmware.api [req-d0bf1a32-a62a-419f-971b-5bfde50dcbcb req-40a715c9-b2d6-4802-8928-626125a75fc8 service nova] Task: {'id': task-1118558, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.987344] env[63531]: DEBUG oslo_vmware.api [None req-71b24bc9-52e0-4981-9ef8-548fbd317f37 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118568, 'name': ReconfigVM_Task, 'duration_secs': 0.153911} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.988663] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-71b24bc9-52e0-4981-9ef8-548fbd317f37 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244796', 'volume_id': 'a34e5b41-b9fc-4c5a-967e-b3e601dde931', 'name': 'volume-a34e5b41-b9fc-4c5a-967e-b3e601dde931', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b7b2b7ac-2321-4cf3-b2ed-4877b893af63', 'attached_at': '', 'detached_at': '', 'volume_id': 'a34e5b41-b9fc-4c5a-967e-b3e601dde931', 'serial': 'a34e5b41-b9fc-4c5a-967e-b3e601dde931'} {{(pid=63531) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1069.071020] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5dff6cb1-9c4d-4e3f-bec7-275c4d5265b6 tempest-ServerRescueNegativeTestJSON-1475768907 tempest-ServerRescueNegativeTestJSON-1475768907-project-member] Lock "4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.119s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1069.128321] env[63531]: DEBUG oslo_vmware.api [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118567, 'name': ReconfigVM_Task, 'duration_secs': 0.74956} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.130821] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 85058146-ba4c-4e9a-bbe7-048ed56167ab] Reconfigured VM instance instance-00000068 to attach disk [datastore2] 85058146-ba4c-4e9a-bbe7-048ed56167ab/85058146-ba4c-4e9a-bbe7-048ed56167ab.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1069.131635] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1ed835e0-6c5a-4162-9c7a-da3f21038a29 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.138876] env[63531]: DEBUG oslo_vmware.api [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 1069.138876] env[63531]: value = "task-1118571" [ 1069.138876] env[63531]: _type = "Task" [ 1069.138876] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.150685] env[63531]: DEBUG oslo_vmware.api [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118571, 'name': Rename_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.190715] env[63531]: DEBUG oslo_concurrency.lockutils [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1069.196691] env[63531]: DEBUG oslo_vmware.api [None req-19dcebb4-f9d8-49dd-bf4f-f27015ebfdab tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118570, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.172697} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.199018] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-19dcebb4-f9d8-49dd-bf4f-f27015ebfdab tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1069.199219] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-19dcebb4-f9d8-49dd-bf4f-f27015ebfdab tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 8f903d83-adbd-4f71-a6f1-037807746248] Deleted contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1069.199416] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-19dcebb4-f9d8-49dd-bf4f-f27015ebfdab tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 8f903d83-adbd-4f71-a6f1-037807746248] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1069.199597] env[63531]: INFO nova.compute.manager [None req-19dcebb4-f9d8-49dd-bf4f-f27015ebfdab tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 8f903d83-adbd-4f71-a6f1-037807746248] Took 0.62 seconds to destroy the instance on the hypervisor. [ 1069.199834] env[63531]: DEBUG oslo.service.loopingcall [None req-19dcebb4-f9d8-49dd-bf4f-f27015ebfdab tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1069.200239] env[63531]: DEBUG nova.compute.manager [-] [instance: 8f903d83-adbd-4f71-a6f1-037807746248] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1069.200325] env[63531]: DEBUG nova.network.neutron [-] [instance: 8f903d83-adbd-4f71-a6f1-037807746248] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1069.209749] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3e97528-e5fe-4b2f-978d-9b7ae70e6b86 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.217764] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86b8d1af-c962-4b4d-bfbb-d69f7d06b0fc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.222737] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Releasing lock "refresh_cache-89f19669-53a9-4425-a3c6-d5f8873b707e" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1069.223044] env[63531]: DEBUG nova.compute.manager [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Instance network_info: |[{"id": "bb17faa5-c291-4bfa-b2fa-c435a815dc3d", "address": "fa:16:3e:2e:a7:41", "network": {"id": "95ed2c7d-3d1a-4736-8e90-104002f5c40b", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-73549344-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9056486e3b640d0bcdb9369ed5d67ad", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04e15990-16e1-4cb2-b0f0-06c362e68c5e", "external-id": "nsx-vlan-transportzone-555", "segmentation_id": 555, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb17faa5-c2", "ovs_interfaceid": "bb17faa5-c291-4bfa-b2fa-c435a815dc3d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1069.223423] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2e:a7:41', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '04e15990-16e1-4cb2-b0f0-06c362e68c5e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bb17faa5-c291-4bfa-b2fa-c435a815dc3d', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1069.230720] env[63531]: DEBUG oslo.service.loopingcall [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1069.254877] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1069.258357] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4634258c-07f3-44c1-9bfd-9a0d084a860f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.274285] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bbe7106-654b-4e9d-899c-804ab924971c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.287494] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1069.287494] env[63531]: value = "task-1118572" [ 1069.287494] env[63531]: _type = "Task" [ 1069.287494] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.289143] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e91120cd-033c-4cd2-9acb-69866105498f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.300372] env[63531]: DEBUG oslo_vmware.api [req-d0bf1a32-a62a-419f-971b-5bfde50dcbcb req-40a715c9-b2d6-4802-8928-626125a75fc8 service nova] Task: {'id': task-1118558, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.310826] env[63531]: DEBUG nova.compute.provider_tree [None req-050cb3b4-5bbb-4627-9fc5-b3033fecceb1 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1069.317843] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118572, 'name': CreateVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.323665] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquiring lock "refresh_cache-18057005-1ff1-4e3d-bccc-8c89755d6db7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1069.323866] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquired lock "refresh_cache-18057005-1ff1-4e3d-bccc-8c89755d6db7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1069.324061] env[63531]: DEBUG nova.network.neutron [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1069.539775] env[63531]: DEBUG nova.compute.manager [req-6d7f0345-f051-4c5e-a521-d6755d6fdaaa req-f1d135ec-72f2-4d7b-8030-0a197a60372c service nova] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Received event network-changed-bb17faa5-c291-4bfa-b2fa-c435a815dc3d {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1069.539775] env[63531]: DEBUG nova.compute.manager [req-6d7f0345-f051-4c5e-a521-d6755d6fdaaa req-f1d135ec-72f2-4d7b-8030-0a197a60372c service nova] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Refreshing instance network info cache due to event network-changed-bb17faa5-c291-4bfa-b2fa-c435a815dc3d. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1069.539775] env[63531]: DEBUG oslo_concurrency.lockutils [req-6d7f0345-f051-4c5e-a521-d6755d6fdaaa req-f1d135ec-72f2-4d7b-8030-0a197a60372c service nova] Acquiring lock "refresh_cache-89f19669-53a9-4425-a3c6-d5f8873b707e" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1069.539775] env[63531]: DEBUG oslo_concurrency.lockutils [req-6d7f0345-f051-4c5e-a521-d6755d6fdaaa req-f1d135ec-72f2-4d7b-8030-0a197a60372c service nova] Acquired lock "refresh_cache-89f19669-53a9-4425-a3c6-d5f8873b707e" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1069.539775] env[63531]: DEBUG nova.network.neutron [req-6d7f0345-f051-4c5e-a521-d6755d6fdaaa req-f1d135ec-72f2-4d7b-8030-0a197a60372c service nova] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Refreshing network info cache for port bb17faa5-c291-4bfa-b2fa-c435a815dc3d {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1069.551950] env[63531]: DEBUG nova.objects.instance [None req-71b24bc9-52e0-4981-9ef8-548fbd317f37 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lazy-loading 'flavor' on Instance uuid b7b2b7ac-2321-4cf3-b2ed-4877b893af63 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1069.649134] env[63531]: DEBUG oslo_vmware.api [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118571, 'name': Rename_Task, 'duration_secs': 0.154805} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.649448] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 85058146-ba4c-4e9a-bbe7-048ed56167ab] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1069.649697] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ced60aed-a8ba-4ccd-b0d0-bb4404d15a59 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.657143] env[63531]: DEBUG oslo_vmware.api [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 1069.657143] env[63531]: value = "task-1118573" [ 1069.657143] env[63531]: _type = "Task" [ 1069.657143] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.674465] env[63531]: DEBUG oslo_vmware.api [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118573, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.791833] env[63531]: DEBUG oslo_vmware.api [req-d0bf1a32-a62a-419f-971b-5bfde50dcbcb req-40a715c9-b2d6-4802-8928-626125a75fc8 service nova] Task: {'id': task-1118558, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.800225] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118572, 'name': CreateVM_Task, 'duration_secs': 0.338107} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.800408] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1069.801201] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1069.801397] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1069.801730] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1069.801986] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b250458e-cb58-4a85-8728-18238103e592 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.809148] env[63531]: DEBUG oslo_vmware.api [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for the task: (returnval){ [ 1069.809148] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]528d0f64-6cb4-c15c-5963-cc065ba234f2" [ 1069.809148] env[63531]: _type = "Task" [ 1069.809148] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.819370] env[63531]: DEBUG nova.scheduler.client.report [None req-050cb3b4-5bbb-4627-9fc5-b3033fecceb1 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1069.822760] env[63531]: DEBUG oslo_vmware.api [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]528d0f64-6cb4-c15c-5963-cc065ba234f2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.999082] env[63531]: DEBUG nova.network.neutron [-] [instance: 8f903d83-adbd-4f71-a6f1-037807746248] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1070.168976] env[63531]: DEBUG oslo_vmware.api [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118573, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.292618] env[63531]: DEBUG oslo_vmware.api [req-d0bf1a32-a62a-419f-971b-5bfde50dcbcb req-40a715c9-b2d6-4802-8928-626125a75fc8 service nova] Task: {'id': task-1118558, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.294788] env[63531]: DEBUG nova.network.neutron [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Updating instance_info_cache with network_info: [{"id": "af2485f6-5ce2-4a83-9d95-7bea85cebff2", "address": "fa:16:3e:f0:7d:92", "network": {"id": "c4f4c994-36ab-42ab-94f8-78f29cbafa50", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-994507698-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "57e9bf4ee0f1463fa67b0f8deff30fdc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "767a3a48-41d4-4a0c-961d-0024837f63bd", "external-id": "nsx-vlan-transportzone-141", "segmentation_id": 141, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaf2485f6-5c", "ovs_interfaceid": "af2485f6-5ce2-4a83-9d95-7bea85cebff2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1070.320073] env[63531]: DEBUG oslo_vmware.api [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]528d0f64-6cb4-c15c-5963-cc065ba234f2, 'name': SearchDatastore_Task, 'duration_secs': 0.010407} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.320387] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1070.320621] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1070.320858] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1070.321017] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1070.321203] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1070.321503] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2d4494ea-5754-4f4d-806c-40c8655d1401 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.323885] env[63531]: DEBUG oslo_concurrency.lockutils [None req-050cb3b4-5bbb-4627-9fc5-b3033fecceb1 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.786s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.325903] env[63531]: DEBUG oslo_concurrency.lockutils [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 1.135s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.335701] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1070.335701] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1070.335836] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-deda2171-e672-472f-9a53-05ca62c8a6da {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.341880] env[63531]: DEBUG oslo_vmware.api [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for the task: (returnval){ [ 1070.341880] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]529a96a3-878b-c3f9-9676-db8522fb4281" [ 1070.341880] env[63531]: _type = "Task" [ 1070.341880] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.346951] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ceb1cec2-fd7f-42f9-9fac-fab6ff52b8e0 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Acquiring lock "b7b2b7ac-2321-4cf3-b2ed-4877b893af63" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.350565] env[63531]: DEBUG oslo_vmware.api [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]529a96a3-878b-c3f9-9676-db8522fb4281, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.351448] env[63531]: INFO nova.scheduler.client.report [None req-050cb3b4-5bbb-4627-9fc5-b3033fecceb1 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Deleted allocations for instance b6db77e6-4447-4b22-93d7-265ff0ada0bd [ 1070.446561] env[63531]: DEBUG nova.network.neutron [req-6d7f0345-f051-4c5e-a521-d6755d6fdaaa req-f1d135ec-72f2-4d7b-8030-0a197a60372c service nova] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Updated VIF entry in instance network info cache for port bb17faa5-c291-4bfa-b2fa-c435a815dc3d. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1070.447317] env[63531]: DEBUG nova.network.neutron [req-6d7f0345-f051-4c5e-a521-d6755d6fdaaa req-f1d135ec-72f2-4d7b-8030-0a197a60372c service nova] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Updating instance_info_cache with network_info: [{"id": "bb17faa5-c291-4bfa-b2fa-c435a815dc3d", "address": "fa:16:3e:2e:a7:41", "network": {"id": "95ed2c7d-3d1a-4736-8e90-104002f5c40b", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-73549344-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9056486e3b640d0bcdb9369ed5d67ad", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04e15990-16e1-4cb2-b0f0-06c362e68c5e", "external-id": "nsx-vlan-transportzone-555", "segmentation_id": 555, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb17faa5-c2", "ovs_interfaceid": "bb17faa5-c291-4bfa-b2fa-c435a815dc3d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1070.501372] env[63531]: INFO nova.compute.manager [-] [instance: 8f903d83-adbd-4f71-a6f1-037807746248] Took 1.30 seconds to deallocate network for instance. [ 1070.559942] env[63531]: DEBUG oslo_concurrency.lockutils [None req-71b24bc9-52e0-4981-9ef8-548fbd317f37 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lock "b7b2b7ac-2321-4cf3-b2ed-4877b893af63" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.272s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.560920] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ceb1cec2-fd7f-42f9-9fac-fab6ff52b8e0 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lock "b7b2b7ac-2321-4cf3-b2ed-4877b893af63" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.214s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.667673] env[63531]: DEBUG oslo_vmware.api [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118573, 'name': PowerOnVM_Task, 'duration_secs': 0.533181} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.667980] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 85058146-ba4c-4e9a-bbe7-048ed56167ab] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1070.668229] env[63531]: INFO nova.compute.manager [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 85058146-ba4c-4e9a-bbe7-048ed56167ab] Took 8.41 seconds to spawn the instance on the hypervisor. [ 1070.668416] env[63531]: DEBUG nova.compute.manager [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 85058146-ba4c-4e9a-bbe7-048ed56167ab] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1070.669312] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-834ae1e3-4630-4175-a06c-a3136b044f7c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.792762] env[63531]: DEBUG oslo_vmware.api [req-d0bf1a32-a62a-419f-971b-5bfde50dcbcb req-40a715c9-b2d6-4802-8928-626125a75fc8 service nova] Task: {'id': task-1118558, 'name': ReconfigVM_Task, 'duration_secs': 5.794117} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.793126] env[63531]: DEBUG oslo_concurrency.lockutils [req-d0bf1a32-a62a-419f-971b-5bfde50dcbcb req-40a715c9-b2d6-4802-8928-626125a75fc8 service nova] Releasing lock "e742d3af-5a45-4274-adfb-c6138bf84d6a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1070.793263] env[63531]: DEBUG nova.virt.vmwareapi.vmops [req-d0bf1a32-a62a-419f-971b-5bfde50dcbcb req-40a715c9-b2d6-4802-8928-626125a75fc8 service nova] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Reconfigured VM to detach interface {{(pid=63531) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 1070.793741] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c2241b8b-55fa-4beb-a8b2-7758c1bb7c17 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "e742d3af-5a45-4274-adfb-c6138bf84d6a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 6.070s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.793958] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c2241b8b-55fa-4beb-a8b2-7758c1bb7c17 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "e742d3af-5a45-4274-adfb-c6138bf84d6a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.794271] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c2241b8b-55fa-4beb-a8b2-7758c1bb7c17 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "e742d3af-5a45-4274-adfb-c6138bf84d6a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.794540] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c2241b8b-55fa-4beb-a8b2-7758c1bb7c17 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "e742d3af-5a45-4274-adfb-c6138bf84d6a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.799143] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Releasing lock "refresh_cache-18057005-1ff1-4e3d-bccc-8c89755d6db7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1070.800336] env[63531]: INFO nova.compute.manager [None req-c2241b8b-55fa-4beb-a8b2-7758c1bb7c17 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Terminating instance [ 1070.802188] env[63531]: DEBUG nova.compute.manager [None req-c2241b8b-55fa-4beb-a8b2-7758c1bb7c17 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1070.802389] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c2241b8b-55fa-4beb-a8b2-7758c1bb7c17 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1070.803201] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e956b0ae-19ba-4e2a-a6fb-5445c4333384 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.811821] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2241b8b-55fa-4beb-a8b2-7758c1bb7c17 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1070.812087] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7bf0d38b-1ea8-43c3-997d-a8f6f15d8e72 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.819138] env[63531]: DEBUG oslo_vmware.api [None req-c2241b8b-55fa-4beb-a8b2-7758c1bb7c17 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for the task: (returnval){ [ 1070.819138] env[63531]: value = "task-1118574" [ 1070.819138] env[63531]: _type = "Task" [ 1070.819138] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.828310] env[63531]: DEBUG oslo_vmware.api [None req-c2241b8b-55fa-4beb-a8b2-7758c1bb7c17 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118574, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.836273] env[63531]: INFO nova.compute.claims [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1070.854443] env[63531]: DEBUG oslo_vmware.api [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]529a96a3-878b-c3f9-9676-db8522fb4281, 'name': SearchDatastore_Task, 'duration_secs': 0.009029} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.859475] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d5bf6a0f-a00f-4883-ab0d-4d397345bb62 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.860583] env[63531]: DEBUG oslo_concurrency.lockutils [None req-050cb3b4-5bbb-4627-9fc5-b3033fecceb1 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "b6db77e6-4447-4b22-93d7-265ff0ada0bd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.098s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.866194] env[63531]: DEBUG oslo_vmware.api [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for the task: (returnval){ [ 1070.866194] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52ca97a9-1d6b-7486-5964-90e8c7766580" [ 1070.866194] env[63531]: _type = "Task" [ 1070.866194] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.876519] env[63531]: DEBUG oslo_vmware.api [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52ca97a9-1d6b-7486-5964-90e8c7766580, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.950193] env[63531]: DEBUG oslo_concurrency.lockutils [req-6d7f0345-f051-4c5e-a521-d6755d6fdaaa req-f1d135ec-72f2-4d7b-8030-0a197a60372c service nova] Releasing lock "refresh_cache-89f19669-53a9-4425-a3c6-d5f8873b707e" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1071.010597] env[63531]: DEBUG oslo_concurrency.lockutils [None req-19dcebb4-f9d8-49dd-bf4f-f27015ebfdab tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1071.064636] env[63531]: INFO nova.compute.manager [None req-ceb1cec2-fd7f-42f9-9fac-fab6ff52b8e0 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Detaching volume 13d7c941-f906-4020-b545-f2bf1bf8a7a0 [ 1071.110598] env[63531]: INFO nova.virt.block_device [None req-ceb1cec2-fd7f-42f9-9fac-fab6ff52b8e0 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Attempting to driver detach volume 13d7c941-f906-4020-b545-f2bf1bf8a7a0 from mountpoint /dev/sdc [ 1071.110711] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-ceb1cec2-fd7f-42f9-9fac-fab6ff52b8e0 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Volume detach. Driver type: vmdk {{(pid=63531) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1071.111350] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-ceb1cec2-fd7f-42f9-9fac-fab6ff52b8e0 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244799', 'volume_id': '13d7c941-f906-4020-b545-f2bf1bf8a7a0', 'name': 'volume-13d7c941-f906-4020-b545-f2bf1bf8a7a0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b7b2b7ac-2321-4cf3-b2ed-4877b893af63', 'attached_at': '', 'detached_at': '', 'volume_id': '13d7c941-f906-4020-b545-f2bf1bf8a7a0', 'serial': '13d7c941-f906-4020-b545-f2bf1bf8a7a0'} {{(pid=63531) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1071.111732] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21aed2dd-93fe-49df-afa5-3f281e24c5c4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.142836] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26983b71-04b2-4012-bd48-292240512c9e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.150685] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1578c57d-cf5a-4cfa-9120-b8fccdb4e0b2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.172365] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb09b6cd-8937-4a16-85f9-497e8a7d3aec {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.192870] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-ceb1cec2-fd7f-42f9-9fac-fab6ff52b8e0 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] The volume has not been displaced from its original location: [datastore2] volume-13d7c941-f906-4020-b545-f2bf1bf8a7a0/volume-13d7c941-f906-4020-b545-f2bf1bf8a7a0.vmdk. No consolidation needed. {{(pid=63531) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1071.198415] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-ceb1cec2-fd7f-42f9-9fac-fab6ff52b8e0 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Reconfiguring VM instance instance-0000005d to detach disk 2002 {{(pid=63531) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1071.201599] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7f1c71d6-41a1-4929-aa53-e7308b2ef8ba {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.213987] env[63531]: INFO nova.compute.manager [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 85058146-ba4c-4e9a-bbe7-048ed56167ab] Took 17.90 seconds to build instance. [ 1071.222741] env[63531]: DEBUG oslo_vmware.api [None req-ceb1cec2-fd7f-42f9-9fac-fab6ff52b8e0 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Waiting for the task: (returnval){ [ 1071.222741] env[63531]: value = "task-1118575" [ 1071.222741] env[63531]: _type = "Task" [ 1071.222741] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.234360] env[63531]: DEBUG oslo_vmware.api [None req-ceb1cec2-fd7f-42f9-9fac-fab6ff52b8e0 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118575, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.307724] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aef58199-b242-402e-91b0-7abaa3d64df3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.316664] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44196f5f-1027-471b-b120-f1e222812f34 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.334487] env[63531]: DEBUG oslo_vmware.api [None req-c2241b8b-55fa-4beb-a8b2-7758c1bb7c17 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118574, 'name': PowerOffVM_Task, 'duration_secs': 0.207456} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.334487] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2241b8b-55fa-4beb-a8b2-7758c1bb7c17 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1071.334487] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c2241b8b-55fa-4beb-a8b2-7758c1bb7c17 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1071.334487] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-be37b310-747d-4f1c-b149-520f709ca3d7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.344736] env[63531]: INFO nova.compute.resource_tracker [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Updating resource usage from migration 063e7061-74ea-430a-93f4-e2b436fafba7 [ 1071.380128] env[63531]: DEBUG oslo_vmware.api [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52ca97a9-1d6b-7486-5964-90e8c7766580, 'name': SearchDatastore_Task, 'duration_secs': 0.01063} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.383748] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1071.383897] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 89f19669-53a9-4425-a3c6-d5f8873b707e/89f19669-53a9-4425-a3c6-d5f8873b707e.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1071.384455] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8fa81521-053d-4221-8d4b-86a784ba4d2e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.394831] env[63531]: DEBUG oslo_vmware.api [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for the task: (returnval){ [ 1071.394831] env[63531]: value = "task-1118577" [ 1071.394831] env[63531]: _type = "Task" [ 1071.394831] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.412548] env[63531]: DEBUG oslo_vmware.api [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118577, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.418109] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c2241b8b-55fa-4beb-a8b2-7758c1bb7c17 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1071.418109] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c2241b8b-55fa-4beb-a8b2-7758c1bb7c17 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1071.418109] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-c2241b8b-55fa-4beb-a8b2-7758c1bb7c17 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Deleting the datastore file [datastore1] e742d3af-5a45-4274-adfb-c6138bf84d6a {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1071.418309] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-628ea21d-57e5-40bf-9389-c98d11a8619a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.430862] env[63531]: DEBUG oslo_vmware.api [None req-c2241b8b-55fa-4beb-a8b2-7758c1bb7c17 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for the task: (returnval){ [ 1071.430862] env[63531]: value = "task-1118578" [ 1071.430862] env[63531]: _type = "Task" [ 1071.430862] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.440986] env[63531]: DEBUG oslo_vmware.api [None req-c2241b8b-55fa-4beb-a8b2-7758c1bb7c17 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118578, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.583032] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3f10148-0b12-4dbd-9ce6-cf0be2b94e8a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.590975] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-866d780a-1694-42d7-95d4-f0339641a475 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.626384] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9854757c-2dc0-4180-bd02-dea0efb0ba02 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.636946] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de3ee8b4-a0ec-4edc-95f0-054766a995b7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.653568] env[63531]: DEBUG nova.compute.provider_tree [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1071.700050] env[63531]: DEBUG nova.compute.manager [req-638a7653-d957-4b37-a493-e2e6b47c14bb req-1328e06a-bf8e-40ca-9f08-e2c8854b5418 service nova] [instance: 8f903d83-adbd-4f71-a6f1-037807746248] Received event network-vif-deleted-0319492b-6961-4fcd-92f1-ddba0968522d {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1071.717602] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f1e68d7b-e1dc-4cb9-b190-13b8976bafbf tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "85058146-ba4c-4e9a-bbe7-048ed56167ab" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.415s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1071.735308] env[63531]: DEBUG oslo_vmware.api [None req-ceb1cec2-fd7f-42f9-9fac-fab6ff52b8e0 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118575, 'name': ReconfigVM_Task, 'duration_secs': 0.33373} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.735626] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-ceb1cec2-fd7f-42f9-9fac-fab6ff52b8e0 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Reconfigured VM instance instance-0000005d to detach disk 2002 {{(pid=63531) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1071.740822] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-390aa1e9-cb8c-4b8b-a71d-8795ecce4e1e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.768485] env[63531]: DEBUG oslo_vmware.api [None req-ceb1cec2-fd7f-42f9-9fac-fab6ff52b8e0 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Waiting for the task: (returnval){ [ 1071.768485] env[63531]: value = "task-1118579" [ 1071.768485] env[63531]: _type = "Task" [ 1071.768485] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.780883] env[63531]: DEBUG oslo_vmware.api [None req-ceb1cec2-fd7f-42f9-9fac-fab6ff52b8e0 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118579, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.907793] env[63531]: DEBUG oslo_vmware.api [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118577, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.484248} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.908109] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 89f19669-53a9-4425-a3c6-d5f8873b707e/89f19669-53a9-4425-a3c6-d5f8873b707e.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1071.908390] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1071.908756] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b394c64e-2370-4450-960b-fe1134ed227c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.942055] env[63531]: DEBUG oslo_concurrency.lockutils [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1071.942111] env[63531]: DEBUG oslo_concurrency.lockutils [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1071.943902] env[63531]: DEBUG oslo_vmware.api [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for the task: (returnval){ [ 1071.943902] env[63531]: value = "task-1118580" [ 1071.943902] env[63531]: _type = "Task" [ 1071.943902] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.950076] env[63531]: DEBUG oslo_vmware.api [None req-c2241b8b-55fa-4beb-a8b2-7758c1bb7c17 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118578, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.4482} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.954122] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-c2241b8b-55fa-4beb-a8b2-7758c1bb7c17 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1071.954354] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c2241b8b-55fa-4beb-a8b2-7758c1bb7c17 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1071.954563] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c2241b8b-55fa-4beb-a8b2-7758c1bb7c17 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1071.954792] env[63531]: INFO nova.compute.manager [None req-c2241b8b-55fa-4beb-a8b2-7758c1bb7c17 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1071.955069] env[63531]: DEBUG oslo.service.loopingcall [None req-c2241b8b-55fa-4beb-a8b2-7758c1bb7c17 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1071.955710] env[63531]: DEBUG nova.compute.manager [-] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1071.955710] env[63531]: DEBUG nova.network.neutron [-] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1071.964062] env[63531]: DEBUG oslo_vmware.api [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118580, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.020454] env[63531]: DEBUG oslo_concurrency.lockutils [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquiring lock "49967454-a4f9-47f6-a2a9-0cd0c29e18d8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.020664] env[63531]: DEBUG oslo_concurrency.lockutils [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "49967454-a4f9-47f6-a2a9-0cd0c29e18d8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1072.073656] env[63531]: DEBUG neutronclient.v2_0.client [-] Error message: {"NeutronError": {"type": "PortNotFound", "message": "Port 4140c945-4976-4856-876d-14507f0a9e17 could not be found.", "detail": ""}} {{(pid=63531) _handle_fault_response /opt/stack/data/venv/lib/python3.10/site-packages/neutronclient/v2_0/client.py:262}} [ 1072.073850] env[63531]: DEBUG nova.network.neutron [-] Unable to show port 4140c945-4976-4856-876d-14507f0a9e17 as it no longer exists. {{(pid=63531) _unbind_ports /opt/stack/nova/nova/network/neutron.py:666}} [ 1072.158275] env[63531]: DEBUG nova.scheduler.client.report [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1072.280499] env[63531]: DEBUG oslo_vmware.api [None req-ceb1cec2-fd7f-42f9-9fac-fab6ff52b8e0 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118579, 'name': ReconfigVM_Task, 'duration_secs': 0.189736} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.280499] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-ceb1cec2-fd7f-42f9-9fac-fab6ff52b8e0 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244799', 'volume_id': '13d7c941-f906-4020-b545-f2bf1bf8a7a0', 'name': 'volume-13d7c941-f906-4020-b545-f2bf1bf8a7a0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b7b2b7ac-2321-4cf3-b2ed-4877b893af63', 'attached_at': '', 'detached_at': '', 'volume_id': '13d7c941-f906-4020-b545-f2bf1bf8a7a0', 'serial': '13d7c941-f906-4020-b545-f2bf1bf8a7a0'} {{(pid=63531) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1072.450821] env[63531]: DEBUG nova.compute.manager [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1072.468557] env[63531]: DEBUG oslo_vmware.api [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118580, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.093129} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.470092] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1072.473885] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbffe9b2-9374-488a-bdea-60041a935ec7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.495823] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Reconfiguring VM instance instance-00000069 to attach disk [datastore1] 89f19669-53a9-4425-a3c6-d5f8873b707e/89f19669-53a9-4425-a3c6-d5f8873b707e.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1072.498110] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-73e9f4d9-c27c-435a-a3bf-fd503b357dee {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.512908] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ebdc669-9ea5-4114-b2f3-c8fe132d6ab0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.537166] env[63531]: DEBUG nova.compute.manager [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1072.543169] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2d4dad5-44cf-4024-b228-e912810dd605 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.546214] env[63531]: DEBUG oslo_vmware.api [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for the task: (returnval){ [ 1072.546214] env[63531]: value = "task-1118581" [ 1072.546214] env[63531]: _type = "Task" [ 1072.546214] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.552304] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Updating instance '18057005-1ff1-4e3d-bccc-8c89755d6db7' progress to 83 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1072.561575] env[63531]: DEBUG oslo_vmware.api [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118581, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.664491] env[63531]: DEBUG oslo_concurrency.lockutils [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.338s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.664709] env[63531]: INFO nova.compute.manager [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Migrating [ 1072.671427] env[63531]: DEBUG oslo_concurrency.lockutils [None req-19dcebb4-f9d8-49dd-bf4f-f27015ebfdab tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.663s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1072.671698] env[63531]: DEBUG nova.objects.instance [None req-19dcebb4-f9d8-49dd-bf4f-f27015ebfdab tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lazy-loading 'resources' on Instance uuid 8f903d83-adbd-4f71-a6f1-037807746248 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1072.827051] env[63531]: DEBUG nova.objects.instance [None req-ceb1cec2-fd7f-42f9-9fac-fab6ff52b8e0 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lazy-loading 'flavor' on Instance uuid b7b2b7ac-2321-4cf3-b2ed-4877b893af63 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1072.975237] env[63531]: DEBUG nova.network.neutron [-] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1072.984067] env[63531]: DEBUG oslo_concurrency.lockutils [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.060144] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1073.060504] env[63531]: DEBUG oslo_vmware.api [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118581, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.061581] env[63531]: DEBUG oslo_concurrency.lockutils [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.061825] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-89448e69-7d0e-4a78-af7f-5d4d044256d5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.070393] env[63531]: DEBUG oslo_vmware.api [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 1073.070393] env[63531]: value = "task-1118582" [ 1073.070393] env[63531]: _type = "Task" [ 1073.070393] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.079154] env[63531]: DEBUG oslo_vmware.api [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118582, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.087029] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f2c55fcf-25ae-49ea-805d-5d27839b6186 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "85058146-ba4c-4e9a-bbe7-048ed56167ab" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.087029] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f2c55fcf-25ae-49ea-805d-5d27839b6186 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "85058146-ba4c-4e9a-bbe7-048ed56167ab" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.087283] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f2c55fcf-25ae-49ea-805d-5d27839b6186 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "85058146-ba4c-4e9a-bbe7-048ed56167ab-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.087534] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f2c55fcf-25ae-49ea-805d-5d27839b6186 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "85058146-ba4c-4e9a-bbe7-048ed56167ab-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.087720] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f2c55fcf-25ae-49ea-805d-5d27839b6186 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "85058146-ba4c-4e9a-bbe7-048ed56167ab-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1073.090179] env[63531]: INFO nova.compute.manager [None req-f2c55fcf-25ae-49ea-805d-5d27839b6186 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 85058146-ba4c-4e9a-bbe7-048ed56167ab] Terminating instance [ 1073.092058] env[63531]: DEBUG nova.compute.manager [None req-f2c55fcf-25ae-49ea-805d-5d27839b6186 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 85058146-ba4c-4e9a-bbe7-048ed56167ab] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1073.092265] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f2c55fcf-25ae-49ea-805d-5d27839b6186 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 85058146-ba4c-4e9a-bbe7-048ed56167ab] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1073.093129] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a9fa15f-c261-4f6b-b281-72289693ee19 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.103164] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2c55fcf-25ae-49ea-805d-5d27839b6186 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 85058146-ba4c-4e9a-bbe7-048ed56167ab] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1073.103410] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-55eacb2b-4e75-4487-ad3f-44c9c97bcf98 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.111901] env[63531]: DEBUG oslo_vmware.api [None req-f2c55fcf-25ae-49ea-805d-5d27839b6186 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 1073.111901] env[63531]: value = "task-1118583" [ 1073.111901] env[63531]: _type = "Task" [ 1073.111901] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.123986] env[63531]: DEBUG oslo_vmware.api [None req-f2c55fcf-25ae-49ea-805d-5d27839b6186 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118583, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.180881] env[63531]: DEBUG oslo_concurrency.lockutils [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "refresh_cache-96c2910b-1e25-4053-9ad9-c155402ef0ac" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1073.181108] env[63531]: DEBUG oslo_concurrency.lockutils [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquired lock "refresh_cache-96c2910b-1e25-4053-9ad9-c155402ef0ac" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1073.181279] env[63531]: DEBUG nova.network.neutron [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1073.387291] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-063a716d-d1a5-4d85-a284-187f27dcf342 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.395414] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78ca134f-caf7-4504-a0ef-9e25e4e83c63 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.425831] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20de830d-b5c9-4db4-9f1c-7c1eba3e3fd1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.433112] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f86f58c4-004c-4ce0-bdd5-83d02d48d0fb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.446571] env[63531]: DEBUG nova.compute.provider_tree [None req-19dcebb4-f9d8-49dd-bf4f-f27015ebfdab tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1073.478868] env[63531]: INFO nova.compute.manager [-] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Took 1.52 seconds to deallocate network for instance. [ 1073.559461] env[63531]: DEBUG oslo_vmware.api [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118581, 'name': ReconfigVM_Task, 'duration_secs': 0.866074} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.559707] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Reconfigured VM instance instance-00000069 to attach disk [datastore1] 89f19669-53a9-4425-a3c6-d5f8873b707e/89f19669-53a9-4425-a3c6-d5f8873b707e.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1073.560346] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0832485c-0066-4d5b-bbbf-2cdd875a9845 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.566944] env[63531]: DEBUG oslo_vmware.api [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for the task: (returnval){ [ 1073.566944] env[63531]: value = "task-1118584" [ 1073.566944] env[63531]: _type = "Task" [ 1073.566944] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.576547] env[63531]: DEBUG oslo_vmware.api [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118584, 'name': Rename_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.581490] env[63531]: DEBUG oslo_vmware.api [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118582, 'name': PowerOnVM_Task, 'duration_secs': 0.433367} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.581729] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1073.581913] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-3e01a33c-5094-40be-9cbd-b473b86942f4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Updating instance '18057005-1ff1-4e3d-bccc-8c89755d6db7' progress to 100 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1073.623795] env[63531]: DEBUG oslo_vmware.api [None req-f2c55fcf-25ae-49ea-805d-5d27839b6186 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118583, 'name': PowerOffVM_Task, 'duration_secs': 0.178787} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.624150] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2c55fcf-25ae-49ea-805d-5d27839b6186 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 85058146-ba4c-4e9a-bbe7-048ed56167ab] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1073.624354] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f2c55fcf-25ae-49ea-805d-5d27839b6186 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 85058146-ba4c-4e9a-bbe7-048ed56167ab] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1073.624624] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-28c7ecdc-b5aa-49ab-a1a9-1b8a84d6d413 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.696398] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f2c55fcf-25ae-49ea-805d-5d27839b6186 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 85058146-ba4c-4e9a-bbe7-048ed56167ab] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1073.696682] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f2c55fcf-25ae-49ea-805d-5d27839b6186 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 85058146-ba4c-4e9a-bbe7-048ed56167ab] Deleting contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1073.696906] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2c55fcf-25ae-49ea-805d-5d27839b6186 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Deleting the datastore file [datastore2] 85058146-ba4c-4e9a-bbe7-048ed56167ab {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1073.697231] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-87abdaed-bda0-46a2-917c-dbf07cb2efe8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.705918] env[63531]: DEBUG oslo_vmware.api [None req-f2c55fcf-25ae-49ea-805d-5d27839b6186 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for the task: (returnval){ [ 1073.705918] env[63531]: value = "task-1118586" [ 1073.705918] env[63531]: _type = "Task" [ 1073.705918] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.715424] env[63531]: DEBUG oslo_vmware.api [None req-f2c55fcf-25ae-49ea-805d-5d27839b6186 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118586, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.727092] env[63531]: DEBUG nova.compute.manager [req-5a8f509d-dbf2-4d2e-9489-cd1c898fdf09 req-81f77511-d30f-4a34-88d6-0d51c6004d9e service nova] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Received event network-vif-deleted-a926b26b-b99b-49f2-b7dd-6ca158ead545 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1073.839808] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ceb1cec2-fd7f-42f9-9fac-fab6ff52b8e0 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lock "b7b2b7ac-2321-4cf3-b2ed-4877b893af63" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.278s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1073.950328] env[63531]: DEBUG nova.scheduler.client.report [None req-19dcebb4-f9d8-49dd-bf4f-f27015ebfdab tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1073.985553] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c2241b8b-55fa-4beb-a8b2-7758c1bb7c17 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1074.077523] env[63531]: DEBUG oslo_vmware.api [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118584, 'name': Rename_Task, 'duration_secs': 0.168965} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.077955] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1074.078108] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b3430946-ddc9-48ab-ad80-40fccac7396a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.085847] env[63531]: DEBUG oslo_vmware.api [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for the task: (returnval){ [ 1074.085847] env[63531]: value = "task-1118587" [ 1074.085847] env[63531]: _type = "Task" [ 1074.085847] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.098074] env[63531]: DEBUG oslo_vmware.api [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118587, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.122211] env[63531]: DEBUG nova.network.neutron [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Updating instance_info_cache with network_info: [{"id": "b731766f-738c-4d77-a00e-98ed6c683e55", "address": "fa:16:3e:d5:26:02", "network": {"id": "0667c176-cfdf-4e43-b152-e00bdc224448", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-243715985-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.217", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "308eabafea9c4188a58a0f1c22074d2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8f580e6-1d86-41ee-9ebe-c531cb9299c6", "external-id": "nsx-vlan-transportzone-150", "segmentation_id": 150, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb731766f-73", "ovs_interfaceid": "b731766f-738c-4d77-a00e-98ed6c683e55", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1074.217111] env[63531]: DEBUG oslo_vmware.api [None req-f2c55fcf-25ae-49ea-805d-5d27839b6186 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Task: {'id': task-1118586, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.137866} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.217396] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2c55fcf-25ae-49ea-805d-5d27839b6186 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1074.217616] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f2c55fcf-25ae-49ea-805d-5d27839b6186 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 85058146-ba4c-4e9a-bbe7-048ed56167ab] Deleted contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1074.217882] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f2c55fcf-25ae-49ea-805d-5d27839b6186 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 85058146-ba4c-4e9a-bbe7-048ed56167ab] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1074.218105] env[63531]: INFO nova.compute.manager [None req-f2c55fcf-25ae-49ea-805d-5d27839b6186 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] [instance: 85058146-ba4c-4e9a-bbe7-048ed56167ab] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1074.218410] env[63531]: DEBUG oslo.service.loopingcall [None req-f2c55fcf-25ae-49ea-805d-5d27839b6186 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1074.218704] env[63531]: DEBUG nova.compute.manager [-] [instance: 85058146-ba4c-4e9a-bbe7-048ed56167ab] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1074.218854] env[63531]: DEBUG nova.network.neutron [-] [instance: 85058146-ba4c-4e9a-bbe7-048ed56167ab] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1074.455702] env[63531]: DEBUG oslo_concurrency.lockutils [None req-19dcebb4-f9d8-49dd-bf4f-f27015ebfdab tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.784s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1074.458350] env[63531]: DEBUG oslo_concurrency.lockutils [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.474s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1074.459937] env[63531]: INFO nova.compute.claims [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1074.495644] env[63531]: INFO nova.scheduler.client.report [None req-19dcebb4-f9d8-49dd-bf4f-f27015ebfdab tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Deleted allocations for instance 8f903d83-adbd-4f71-a6f1-037807746248 [ 1074.601164] env[63531]: DEBUG oslo_vmware.api [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118587, 'name': PowerOnVM_Task, 'duration_secs': 0.510088} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.601445] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1074.601702] env[63531]: INFO nova.compute.manager [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Took 7.77 seconds to spawn the instance on the hypervisor. [ 1074.601896] env[63531]: DEBUG nova.compute.manager [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1074.602725] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-358a220c-efb0-4ccd-abea-19b907b68113 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.624978] env[63531]: DEBUG oslo_concurrency.lockutils [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Releasing lock "refresh_cache-96c2910b-1e25-4053-9ad9-c155402ef0ac" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1074.794215] env[63531]: DEBUG oslo_concurrency.lockutils [None req-cf55f1a9-a326-424b-abfc-9e7077a42e78 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Acquiring lock "b7b2b7ac-2321-4cf3-b2ed-4877b893af63" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1074.794561] env[63531]: DEBUG oslo_concurrency.lockutils [None req-cf55f1a9-a326-424b-abfc-9e7077a42e78 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lock "b7b2b7ac-2321-4cf3-b2ed-4877b893af63" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1074.794822] env[63531]: DEBUG oslo_concurrency.lockutils [None req-cf55f1a9-a326-424b-abfc-9e7077a42e78 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Acquiring lock "b7b2b7ac-2321-4cf3-b2ed-4877b893af63-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1074.795466] env[63531]: DEBUG oslo_concurrency.lockutils [None req-cf55f1a9-a326-424b-abfc-9e7077a42e78 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lock "b7b2b7ac-2321-4cf3-b2ed-4877b893af63-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1074.795745] env[63531]: DEBUG oslo_concurrency.lockutils [None req-cf55f1a9-a326-424b-abfc-9e7077a42e78 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lock "b7b2b7ac-2321-4cf3-b2ed-4877b893af63-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1074.797958] env[63531]: INFO nova.compute.manager [None req-cf55f1a9-a326-424b-abfc-9e7077a42e78 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Terminating instance [ 1074.799774] env[63531]: DEBUG nova.compute.manager [None req-cf55f1a9-a326-424b-abfc-9e7077a42e78 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1074.799975] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-cf55f1a9-a326-424b-abfc-9e7077a42e78 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1074.800948] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86bff733-09ed-420a-bdcf-edcf86ca8a83 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.810037] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf55f1a9-a326-424b-abfc-9e7077a42e78 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1074.810329] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-489de8ca-ce1a-4e0c-8426-5c49621aeb25 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.818247] env[63531]: DEBUG oslo_vmware.api [None req-cf55f1a9-a326-424b-abfc-9e7077a42e78 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Waiting for the task: (returnval){ [ 1074.818247] env[63531]: value = "task-1118588" [ 1074.818247] env[63531]: _type = "Task" [ 1074.818247] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.827135] env[63531]: DEBUG oslo_vmware.api [None req-cf55f1a9-a326-424b-abfc-9e7077a42e78 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118588, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.988643] env[63531]: DEBUG nova.network.neutron [-] [instance: 85058146-ba4c-4e9a-bbe7-048ed56167ab] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1075.004535] env[63531]: DEBUG oslo_concurrency.lockutils [None req-19dcebb4-f9d8-49dd-bf4f-f27015ebfdab tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "8f903d83-adbd-4f71-a6f1-037807746248" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.430s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1075.126834] env[63531]: INFO nova.compute.manager [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Took 20.22 seconds to build instance. [ 1075.332282] env[63531]: DEBUG oslo_vmware.api [None req-cf55f1a9-a326-424b-abfc-9e7077a42e78 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118588, 'name': PowerOffVM_Task, 'duration_secs': 0.246964} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.333278] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf55f1a9-a326-424b-abfc-9e7077a42e78 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1075.333544] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-cf55f1a9-a326-424b-abfc-9e7077a42e78 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1075.333877] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7e147b01-b406-4858-b3f6-3a96bd916cf1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.408045] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-cf55f1a9-a326-424b-abfc-9e7077a42e78 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1075.408313] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-cf55f1a9-a326-424b-abfc-9e7077a42e78 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Deleting contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1075.408605] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-cf55f1a9-a326-424b-abfc-9e7077a42e78 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Deleting the datastore file [datastore2] b7b2b7ac-2321-4cf3-b2ed-4877b893af63 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1075.408944] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-36473518-52b9-41a2-9786-f6a617d31a00 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.416348] env[63531]: DEBUG oslo_vmware.api [None req-cf55f1a9-a326-424b-abfc-9e7077a42e78 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Waiting for the task: (returnval){ [ 1075.416348] env[63531]: value = "task-1118590" [ 1075.416348] env[63531]: _type = "Task" [ 1075.416348] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.424714] env[63531]: DEBUG oslo_vmware.api [None req-cf55f1a9-a326-424b-abfc-9e7077a42e78 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118590, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.492410] env[63531]: INFO nova.compute.manager [-] [instance: 85058146-ba4c-4e9a-bbe7-048ed56167ab] Took 1.27 seconds to deallocate network for instance. [ 1075.632072] env[63531]: DEBUG oslo_concurrency.lockutils [None req-b4524311-b75a-4221-bb1d-ff34b194d9bc tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "89f19669-53a9-4425-a3c6-d5f8873b707e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.733s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1075.646222] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2470402e-1eef-4ea5-8e92-1a0f64b42a16 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.654064] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-509e98b9-ac10-41a0-9a76-eb233dca23a9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.687213] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36f4e16e-2159-41da-8560-97279e5641d1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.695759] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7440ad24-f669-40fd-a442-a54070226d3c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.709773] env[63531]: DEBUG nova.compute.provider_tree [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1075.752381] env[63531]: DEBUG nova.compute.manager [req-489c24f6-55bf-4eb8-8ea9-9befc6047fa0 req-8d4e1c5a-23d5-4114-aaf8-abe7738b123e service nova] [instance: 85058146-ba4c-4e9a-bbe7-048ed56167ab] Received event network-vif-deleted-9216e469-ea4e-4e8a-9f74-e9310da25110 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1075.938484] env[63531]: DEBUG oslo_vmware.api [None req-cf55f1a9-a326-424b-abfc-9e7077a42e78 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Task: {'id': task-1118590, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.132985} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.940012] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-cf55f1a9-a326-424b-abfc-9e7077a42e78 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1075.940251] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-cf55f1a9-a326-424b-abfc-9e7077a42e78 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Deleted contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1075.940445] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-cf55f1a9-a326-424b-abfc-9e7077a42e78 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1075.940626] env[63531]: INFO nova.compute.manager [None req-cf55f1a9-a326-424b-abfc-9e7077a42e78 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1075.940869] env[63531]: DEBUG oslo.service.loopingcall [None req-cf55f1a9-a326-424b-abfc-9e7077a42e78 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1075.944079] env[63531]: DEBUG nova.compute.manager [-] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1075.944189] env[63531]: DEBUG nova.network.neutron [-] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1075.946225] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "148c6ecd-354e-4076-b52b-c13a2119da7e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1075.946528] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "148c6ecd-354e-4076-b52b-c13a2119da7e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1075.999572] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f2c55fcf-25ae-49ea-805d-5d27839b6186 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1076.004213] env[63531]: DEBUG nova.network.neutron [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Port af2485f6-5ce2-4a83-9d95-7bea85cebff2 binding to destination host cpu-1 is already ACTIVE {{(pid=63531) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1076.004488] env[63531]: DEBUG oslo_concurrency.lockutils [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquiring lock "refresh_cache-18057005-1ff1-4e3d-bccc-8c89755d6db7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1076.005211] env[63531]: DEBUG oslo_concurrency.lockutils [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquired lock "refresh_cache-18057005-1ff1-4e3d-bccc-8c89755d6db7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1076.005614] env[63531]: DEBUG nova.network.neutron [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1076.144126] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c310a09d-8f9e-4aeb-91e3-334f87a31364 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.167369] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Updating instance '96c2910b-1e25-4053-9ad9-c155402ef0ac' progress to 0 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1076.213113] env[63531]: DEBUG nova.scheduler.client.report [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1076.451749] env[63531]: DEBUG nova.compute.manager [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1076.676274] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1076.676274] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5f8fce93-fe60-47d5-bff3-fcd5191c454c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.687529] env[63531]: DEBUG oslo_vmware.api [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 1076.687529] env[63531]: value = "task-1118591" [ 1076.687529] env[63531]: _type = "Task" [ 1076.687529] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.701567] env[63531]: DEBUG oslo_vmware.api [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118591, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.718927] env[63531]: DEBUG oslo_concurrency.lockutils [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.261s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1076.719719] env[63531]: DEBUG nova.compute.manager [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1076.726631] env[63531]: DEBUG oslo_concurrency.lockutils [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.665s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1076.728650] env[63531]: INFO nova.compute.claims [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1076.735614] env[63531]: DEBUG nova.network.neutron [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Updating instance_info_cache with network_info: [{"id": "af2485f6-5ce2-4a83-9d95-7bea85cebff2", "address": "fa:16:3e:f0:7d:92", "network": {"id": "c4f4c994-36ab-42ab-94f8-78f29cbafa50", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-994507698-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "57e9bf4ee0f1463fa67b0f8deff30fdc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "767a3a48-41d4-4a0c-961d-0024837f63bd", "external-id": "nsx-vlan-transportzone-141", "segmentation_id": 141, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaf2485f6-5c", "ovs_interfaceid": "af2485f6-5ce2-4a83-9d95-7bea85cebff2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1076.974376] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1076.977149] env[63531]: DEBUG nova.network.neutron [-] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1077.197433] env[63531]: DEBUG oslo_vmware.api [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118591, 'name': PowerOffVM_Task, 'duration_secs': 0.224706} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.197788] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1077.197890] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Updating instance '96c2910b-1e25-4053-9ad9-c155402ef0ac' progress to 17 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1077.228549] env[63531]: DEBUG nova.compute.utils [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1077.230146] env[63531]: DEBUG nova.compute.manager [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1077.230316] env[63531]: DEBUG nova.network.neutron [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1077.237940] env[63531]: DEBUG oslo_concurrency.lockutils [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Releasing lock "refresh_cache-18057005-1ff1-4e3d-bccc-8c89755d6db7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1077.280578] env[63531]: DEBUG nova.policy [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7140e12106504b3e95ff49786361771b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '63900312a30b4942854f552758c039cf', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 1077.479924] env[63531]: INFO nova.compute.manager [-] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Took 1.54 seconds to deallocate network for instance. [ 1077.535338] env[63531]: DEBUG nova.network.neutron [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] Successfully created port: 6ae9a4f8-494f-4ba5-997a-0dcf67e4320a {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1077.705058] env[63531]: DEBUG nova.virt.hardware [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1077.705350] env[63531]: DEBUG nova.virt.hardware [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1077.705518] env[63531]: DEBUG nova.virt.hardware [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1077.705807] env[63531]: DEBUG nova.virt.hardware [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1077.706174] env[63531]: DEBUG nova.virt.hardware [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1077.706388] env[63531]: DEBUG nova.virt.hardware [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1077.706742] env[63531]: DEBUG nova.virt.hardware [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1077.706967] env[63531]: DEBUG nova.virt.hardware [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1077.707223] env[63531]: DEBUG nova.virt.hardware [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1077.707438] env[63531]: DEBUG nova.virt.hardware [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1077.707689] env[63531]: DEBUG nova.virt.hardware [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1077.713342] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-971b59ee-8423-4153-86ae-020908cd51ce {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.731842] env[63531]: DEBUG oslo_vmware.api [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 1077.731842] env[63531]: value = "task-1118592" [ 1077.731842] env[63531]: _type = "Task" [ 1077.731842] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.737153] env[63531]: DEBUG nova.compute.manager [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1077.747021] env[63531]: DEBUG nova.compute.manager [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=63531) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 1077.755984] env[63531]: DEBUG oslo_vmware.api [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118592, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.798341] env[63531]: DEBUG nova.compute.manager [req-d9b5dd68-3979-4324-83f0-6858500df2b1 req-6dc209e2-764f-4fd9-b469-35e229d467f3 service nova] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Received event network-changed-bb17faa5-c291-4bfa-b2fa-c435a815dc3d {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1077.798626] env[63531]: DEBUG nova.compute.manager [req-d9b5dd68-3979-4324-83f0-6858500df2b1 req-6dc209e2-764f-4fd9-b469-35e229d467f3 service nova] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Refreshing instance network info cache due to event network-changed-bb17faa5-c291-4bfa-b2fa-c435a815dc3d. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1077.798857] env[63531]: DEBUG oslo_concurrency.lockutils [req-d9b5dd68-3979-4324-83f0-6858500df2b1 req-6dc209e2-764f-4fd9-b469-35e229d467f3 service nova] Acquiring lock "refresh_cache-89f19669-53a9-4425-a3c6-d5f8873b707e" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1077.798995] env[63531]: DEBUG oslo_concurrency.lockutils [req-d9b5dd68-3979-4324-83f0-6858500df2b1 req-6dc209e2-764f-4fd9-b469-35e229d467f3 service nova] Acquired lock "refresh_cache-89f19669-53a9-4425-a3c6-d5f8873b707e" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1077.799224] env[63531]: DEBUG nova.network.neutron [req-d9b5dd68-3979-4324-83f0-6858500df2b1 req-6dc209e2-764f-4fd9-b469-35e229d467f3 service nova] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Refreshing network info cache for port bb17faa5-c291-4bfa-b2fa-c435a815dc3d {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1077.944755] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9b34e45-085a-47a9-b2cc-51d26613e1f8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.954635] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7edbbd9-4839-40bc-95b0-019113a91f97 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.988186] env[63531]: DEBUG oslo_concurrency.lockutils [None req-cf55f1a9-a326-424b-abfc-9e7077a42e78 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1077.989272] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1383fc1a-3085-4a76-a8ad-0fc3d0233267 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.998500] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d36e6e5-52ec-48e5-9d89-8a3febb5d96d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.015820] env[63531]: DEBUG nova.compute.provider_tree [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1078.242220] env[63531]: DEBUG oslo_vmware.api [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118592, 'name': ReconfigVM_Task, 'duration_secs': 0.354604} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.242557] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Updating instance '96c2910b-1e25-4053-9ad9-c155402ef0ac' progress to 33 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1078.513542] env[63531]: DEBUG nova.network.neutron [req-d9b5dd68-3979-4324-83f0-6858500df2b1 req-6dc209e2-764f-4fd9-b469-35e229d467f3 service nova] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Updated VIF entry in instance network info cache for port bb17faa5-c291-4bfa-b2fa-c435a815dc3d. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1078.513926] env[63531]: DEBUG nova.network.neutron [req-d9b5dd68-3979-4324-83f0-6858500df2b1 req-6dc209e2-764f-4fd9-b469-35e229d467f3 service nova] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Updating instance_info_cache with network_info: [{"id": "bb17faa5-c291-4bfa-b2fa-c435a815dc3d", "address": "fa:16:3e:2e:a7:41", "network": {"id": "95ed2c7d-3d1a-4736-8e90-104002f5c40b", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-73549344-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.245", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9056486e3b640d0bcdb9369ed5d67ad", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04e15990-16e1-4cb2-b0f0-06c362e68c5e", "external-id": "nsx-vlan-transportzone-555", "segmentation_id": 555, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb17faa5-c2", "ovs_interfaceid": "bb17faa5-c291-4bfa-b2fa-c435a815dc3d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1078.521372] env[63531]: DEBUG nova.scheduler.client.report [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1078.748957] env[63531]: DEBUG nova.virt.hardware [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1078.749339] env[63531]: DEBUG nova.virt.hardware [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1078.749548] env[63531]: DEBUG nova.virt.hardware [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1078.749773] env[63531]: DEBUG nova.virt.hardware [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1078.749958] env[63531]: DEBUG nova.virt.hardware [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1078.750169] env[63531]: DEBUG nova.virt.hardware [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1078.750435] env[63531]: DEBUG nova.virt.hardware [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1078.750636] env[63531]: DEBUG nova.virt.hardware [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1078.750842] env[63531]: DEBUG nova.virt.hardware [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1078.751051] env[63531]: DEBUG nova.virt.hardware [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1078.751309] env[63531]: DEBUG nova.virt.hardware [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1078.756885] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Reconfiguring VM instance instance-00000063 to detach disk 2000 {{(pid=63531) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1078.757487] env[63531]: DEBUG nova.compute.manager [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1078.759474] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6905ca37-8302-46b4-9b70-d8e3b25e8184 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.779575] env[63531]: DEBUG oslo_vmware.api [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 1078.779575] env[63531]: value = "task-1118593" [ 1078.779575] env[63531]: _type = "Task" [ 1078.779575] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.791672] env[63531]: DEBUG oslo_vmware.api [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118593, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.796961] env[63531]: DEBUG nova.virt.hardware [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1078.797257] env[63531]: DEBUG nova.virt.hardware [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1078.797474] env[63531]: DEBUG nova.virt.hardware [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1078.797681] env[63531]: DEBUG nova.virt.hardware [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1078.797835] env[63531]: DEBUG nova.virt.hardware [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1078.797987] env[63531]: DEBUG nova.virt.hardware [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1078.798238] env[63531]: DEBUG nova.virt.hardware [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1078.798404] env[63531]: DEBUG nova.virt.hardware [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1078.798573] env[63531]: DEBUG nova.virt.hardware [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1078.798735] env[63531]: DEBUG nova.virt.hardware [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1078.798908] env[63531]: DEBUG nova.virt.hardware [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1078.799759] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe8f1759-9f05-441a-8fd5-844be8678341 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.807902] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24f2139c-4317-40b0-bf64-a0065c8e3564 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.873867] env[63531]: DEBUG oslo_concurrency.lockutils [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1079.017743] env[63531]: DEBUG oslo_concurrency.lockutils [req-d9b5dd68-3979-4324-83f0-6858500df2b1 req-6dc209e2-764f-4fd9-b469-35e229d467f3 service nova] Releasing lock "refresh_cache-89f19669-53a9-4425-a3c6-d5f8873b707e" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1079.017743] env[63531]: DEBUG nova.compute.manager [req-d9b5dd68-3979-4324-83f0-6858500df2b1 req-6dc209e2-764f-4fd9-b469-35e229d467f3 service nova] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Received event network-vif-deleted-40adf4b3-527a-4414-b2dc-a24a39b897c6 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1079.027151] env[63531]: DEBUG oslo_concurrency.lockutils [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.300s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1079.027151] env[63531]: DEBUG nova.compute.manager [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1079.029580] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c2241b8b-55fa-4beb-a8b2-7758c1bb7c17 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.044s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1079.029791] env[63531]: DEBUG nova.objects.instance [None req-c2241b8b-55fa-4beb-a8b2-7758c1bb7c17 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lazy-loading 'resources' on Instance uuid e742d3af-5a45-4274-adfb-c6138bf84d6a {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1079.050702] env[63531]: DEBUG nova.network.neutron [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] Successfully updated port: 6ae9a4f8-494f-4ba5-997a-0dcf67e4320a {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1079.290392] env[63531]: DEBUG oslo_vmware.api [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118593, 'name': ReconfigVM_Task, 'duration_secs': 0.188612} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.290717] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Reconfigured VM instance instance-00000063 to detach disk 2000 {{(pid=63531) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1079.291517] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84cd904a-c9d1-4a57-97a0-9eb7bbdf6705 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.313311] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] 96c2910b-1e25-4053-9ad9-c155402ef0ac/96c2910b-1e25-4053-9ad9-c155402ef0ac.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1079.313584] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3d5deab8-e6ce-4aa0-abc7-188268c1d110 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.331590] env[63531]: DEBUG oslo_vmware.api [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 1079.331590] env[63531]: value = "task-1118594" [ 1079.331590] env[63531]: _type = "Task" [ 1079.331590] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.339438] env[63531]: DEBUG oslo_vmware.api [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118594, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.532504] env[63531]: DEBUG nova.compute.utils [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1079.534068] env[63531]: DEBUG nova.compute.manager [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1079.534260] env[63531]: DEBUG nova.network.neutron [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1079.554570] env[63531]: DEBUG oslo_concurrency.lockutils [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "refresh_cache-b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1079.554712] env[63531]: DEBUG oslo_concurrency.lockutils [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquired lock "refresh_cache-b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1079.554925] env[63531]: DEBUG nova.network.neutron [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1079.586234] env[63531]: DEBUG nova.policy [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0a501b2c64c04dcda3596c8a1bd3fa0d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'abdae960d43b47bc95ab41f265d2c85b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 1079.699146] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3c32ec6-080a-4406-94fd-a180bd137a1a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.707642] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fda81e66-1514-4bfe-a915-33f849b2f283 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.738905] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ac6c9ac-2da6-4f65-be0b-3a6c2f9ce3fe {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.746554] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-677e923a-8d94-4000-9223-601869662b6c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.759613] env[63531]: DEBUG nova.compute.provider_tree [None req-c2241b8b-55fa-4beb-a8b2-7758c1bb7c17 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1079.821455] env[63531]: DEBUG nova.compute.manager [req-effa0739-b90b-4f14-a1d9-69af34ae5984 req-f87d389d-1f95-4684-b573-d63f775770ce service nova] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] Received event network-vif-plugged-6ae9a4f8-494f-4ba5-997a-0dcf67e4320a {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1079.821576] env[63531]: DEBUG oslo_concurrency.lockutils [req-effa0739-b90b-4f14-a1d9-69af34ae5984 req-f87d389d-1f95-4684-b573-d63f775770ce service nova] Acquiring lock "b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1079.821795] env[63531]: DEBUG oslo_concurrency.lockutils [req-effa0739-b90b-4f14-a1d9-69af34ae5984 req-f87d389d-1f95-4684-b573-d63f775770ce service nova] Lock "b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1079.821963] env[63531]: DEBUG oslo_concurrency.lockutils [req-effa0739-b90b-4f14-a1d9-69af34ae5984 req-f87d389d-1f95-4684-b573-d63f775770ce service nova] Lock "b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1079.822317] env[63531]: DEBUG nova.compute.manager [req-effa0739-b90b-4f14-a1d9-69af34ae5984 req-f87d389d-1f95-4684-b573-d63f775770ce service nova] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] No waiting events found dispatching network-vif-plugged-6ae9a4f8-494f-4ba5-997a-0dcf67e4320a {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1079.822523] env[63531]: WARNING nova.compute.manager [req-effa0739-b90b-4f14-a1d9-69af34ae5984 req-f87d389d-1f95-4684-b573-d63f775770ce service nova] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] Received unexpected event network-vif-plugged-6ae9a4f8-494f-4ba5-997a-0dcf67e4320a for instance with vm_state building and task_state spawning. [ 1079.822695] env[63531]: DEBUG nova.compute.manager [req-effa0739-b90b-4f14-a1d9-69af34ae5984 req-f87d389d-1f95-4684-b573-d63f775770ce service nova] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] Received event network-changed-6ae9a4f8-494f-4ba5-997a-0dcf67e4320a {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1079.822856] env[63531]: DEBUG nova.compute.manager [req-effa0739-b90b-4f14-a1d9-69af34ae5984 req-f87d389d-1f95-4684-b573-d63f775770ce service nova] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] Refreshing instance network info cache due to event network-changed-6ae9a4f8-494f-4ba5-997a-0dcf67e4320a. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1079.823043] env[63531]: DEBUG oslo_concurrency.lockutils [req-effa0739-b90b-4f14-a1d9-69af34ae5984 req-f87d389d-1f95-4684-b573-d63f775770ce service nova] Acquiring lock "refresh_cache-b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1079.842356] env[63531]: DEBUG oslo_vmware.api [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118594, 'name': ReconfigVM_Task, 'duration_secs': 0.260491} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.842753] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Reconfigured VM instance instance-00000063 to attach disk [datastore1] 96c2910b-1e25-4053-9ad9-c155402ef0ac/96c2910b-1e25-4053-9ad9-c155402ef0ac.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1079.843054] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Updating instance '96c2910b-1e25-4053-9ad9-c155402ef0ac' progress to 50 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1079.873217] env[63531]: DEBUG nova.network.neutron [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Successfully created port: 99a45e7c-1ddb-4c35-ac69-b2c33f976019 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1080.039732] env[63531]: DEBUG nova.compute.manager [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1080.099303] env[63531]: DEBUG nova.network.neutron [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1080.239566] env[63531]: DEBUG nova.network.neutron [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] Updating instance_info_cache with network_info: [{"id": "6ae9a4f8-494f-4ba5-997a-0dcf67e4320a", "address": "fa:16:3e:01:fe:a1", "network": {"id": "80641cc0-680b-40ff-8d1b-f5a5f736a88f", "bridge": "br-int", "label": "tempest-ServersTestJSON-666288463-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "63900312a30b4942854f552758c039cf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "26472e27-9835-4f87-ab7f-ca24dfee4e83", "external-id": "nsx-vlan-transportzone-335", "segmentation_id": 335, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ae9a4f8-49", "ovs_interfaceid": "6ae9a4f8-494f-4ba5-997a-0dcf67e4320a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1080.262612] env[63531]: DEBUG nova.scheduler.client.report [None req-c2241b8b-55fa-4beb-a8b2-7758c1bb7c17 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1080.350118] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9847b331-4476-4e93-b4c4-b51bce995e63 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.372378] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8db90944-0b97-4726-b4ca-c224bd9bc655 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.390497] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Updating instance '96c2910b-1e25-4053-9ad9-c155402ef0ac' progress to 67 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1080.545383] env[63531]: INFO nova.virt.block_device [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Booting with volume 422ea428-1dc4-40f2-8151-a5e18a596f9b at /dev/sda [ 1080.582633] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cc76e460-c1a9-40ea-9570-23c069cde2bb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.592592] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4e1ac23-3646-448e-ac45-67c25cb2ecb5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.620845] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e7e03baa-0296-479d-93e7-cfcb4d248dac {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.629088] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00eae454-1c79-45d3-9112-6666d73b4ed3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.659390] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00666d39-2392-4bfe-a942-9a0b818df9f3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.666075] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c3f3cfd-703c-40c5-9590-870f9ebc6b06 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.679761] env[63531]: DEBUG nova.virt.block_device [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Updating existing volume attachment record: 4db59766-5c8c-4c09-bb08-c72e703547ba {{(pid=63531) _volume_attach /opt/stack/nova/nova/virt/block_device.py:679}} [ 1080.741875] env[63531]: DEBUG oslo_concurrency.lockutils [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Releasing lock "refresh_cache-b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1080.742269] env[63531]: DEBUG nova.compute.manager [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] Instance network_info: |[{"id": "6ae9a4f8-494f-4ba5-997a-0dcf67e4320a", "address": "fa:16:3e:01:fe:a1", "network": {"id": "80641cc0-680b-40ff-8d1b-f5a5f736a88f", "bridge": "br-int", "label": "tempest-ServersTestJSON-666288463-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "63900312a30b4942854f552758c039cf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "26472e27-9835-4f87-ab7f-ca24dfee4e83", "external-id": "nsx-vlan-transportzone-335", "segmentation_id": 335, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ae9a4f8-49", "ovs_interfaceid": "6ae9a4f8-494f-4ba5-997a-0dcf67e4320a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1080.742601] env[63531]: DEBUG oslo_concurrency.lockutils [req-effa0739-b90b-4f14-a1d9-69af34ae5984 req-f87d389d-1f95-4684-b573-d63f775770ce service nova] Acquired lock "refresh_cache-b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1080.742789] env[63531]: DEBUG nova.network.neutron [req-effa0739-b90b-4f14-a1d9-69af34ae5984 req-f87d389d-1f95-4684-b573-d63f775770ce service nova] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] Refreshing network info cache for port 6ae9a4f8-494f-4ba5-997a-0dcf67e4320a {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1080.744075] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:01:fe:a1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '26472e27-9835-4f87-ab7f-ca24dfee4e83', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6ae9a4f8-494f-4ba5-997a-0dcf67e4320a', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1080.752089] env[63531]: DEBUG oslo.service.loopingcall [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1080.755154] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1080.755661] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-15b6161f-0c9a-406e-99ec-f7aa9acce24d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.772178] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c2241b8b-55fa-4beb-a8b2-7758c1bb7c17 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.743s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1080.774165] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f2c55fcf-25ae-49ea-805d-5d27839b6186 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.775s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1080.774395] env[63531]: DEBUG nova.objects.instance [None req-f2c55fcf-25ae-49ea-805d-5d27839b6186 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lazy-loading 'resources' on Instance uuid 85058146-ba4c-4e9a-bbe7-048ed56167ab {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1080.782477] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1080.782477] env[63531]: value = "task-1118595" [ 1080.782477] env[63531]: _type = "Task" [ 1080.782477] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.791909] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118595, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.797404] env[63531]: INFO nova.scheduler.client.report [None req-c2241b8b-55fa-4beb-a8b2-7758c1bb7c17 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Deleted allocations for instance e742d3af-5a45-4274-adfb-c6138bf84d6a [ 1080.942467] env[63531]: DEBUG nova.network.neutron [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Port b731766f-738c-4d77-a00e-98ed6c683e55 binding to destination host cpu-1 is already ACTIVE {{(pid=63531) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1080.982084] env[63531]: DEBUG nova.network.neutron [req-effa0739-b90b-4f14-a1d9-69af34ae5984 req-f87d389d-1f95-4684-b573-d63f775770ce service nova] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] Updated VIF entry in instance network info cache for port 6ae9a4f8-494f-4ba5-997a-0dcf67e4320a. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1080.982496] env[63531]: DEBUG nova.network.neutron [req-effa0739-b90b-4f14-a1d9-69af34ae5984 req-f87d389d-1f95-4684-b573-d63f775770ce service nova] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] Updating instance_info_cache with network_info: [{"id": "6ae9a4f8-494f-4ba5-997a-0dcf67e4320a", "address": "fa:16:3e:01:fe:a1", "network": {"id": "80641cc0-680b-40ff-8d1b-f5a5f736a88f", "bridge": "br-int", "label": "tempest-ServersTestJSON-666288463-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "63900312a30b4942854f552758c039cf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "26472e27-9835-4f87-ab7f-ca24dfee4e83", "external-id": "nsx-vlan-transportzone-335", "segmentation_id": 335, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ae9a4f8-49", "ovs_interfaceid": "6ae9a4f8-494f-4ba5-997a-0dcf67e4320a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1081.026208] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1081.026489] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1081.026657] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Starting heal instance info cache {{(pid=63531) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1081.026790] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Rebuilding the list of instances to heal {{(pid=63531) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1081.293732] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118595, 'name': CreateVM_Task, 'duration_secs': 0.349948} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.293939] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1081.297244] env[63531]: DEBUG oslo_concurrency.lockutils [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1081.297244] env[63531]: DEBUG oslo_concurrency.lockutils [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1081.297426] env[63531]: DEBUG oslo_concurrency.lockutils [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1081.298158] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5edfe978-08d8-4753-ac25-eefd7ee62596 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.303234] env[63531]: DEBUG oslo_vmware.api [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 1081.303234] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52ccf061-bf44-4c4a-920e-bd5df919d062" [ 1081.303234] env[63531]: _type = "Task" [ 1081.303234] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.310326] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c2241b8b-55fa-4beb-a8b2-7758c1bb7c17 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "e742d3af-5a45-4274-adfb-c6138bf84d6a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.515s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1081.310326] env[63531]: DEBUG oslo_concurrency.lockutils [req-12df086a-aaee-4fc0-9c79-a53119d48317 req-8bea5b09-c55f-421c-be42-47aed0180cfb service nova] Acquired lock "e742d3af-5a45-4274-adfb-c6138bf84d6a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1081.314025] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-693dbe89-f0f7-458a-957c-66242d2d4529 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.318988] env[63531]: DEBUG oslo_vmware.api [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52ccf061-bf44-4c4a-920e-bd5df919d062, 'name': SearchDatastore_Task, 'duration_secs': 0.009936} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.318988] env[63531]: DEBUG oslo_concurrency.lockutils [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1081.318988] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1081.318988] env[63531]: DEBUG oslo_concurrency.lockutils [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1081.318988] env[63531]: DEBUG oslo_concurrency.lockutils [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1081.319513] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1081.319513] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ea91d9cd-9d50-4e62-9042-93e19cd7240f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.326212] env[63531]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 1081.326373] env[63531]: DEBUG oslo_vmware.api [-] Fault list: [ManagedObjectNotFound] {{(pid=63531) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 1081.327479] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-16e2d885-6b51-4295-89a5-e1a94080f0c5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.329872] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1081.330071] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1081.331105] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-107f1606-6c4d-4bbd-97d2-02bd234a8e7f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.337825] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbb63165-1802-4a44-aa72-bdef36d3f52c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.355270] env[63531]: DEBUG oslo_vmware.api [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 1081.355270] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52c4c044-9be0-945b-96bb-4cfc7ef17d0c" [ 1081.355270] env[63531]: _type = "Task" [ 1081.355270] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.365770] env[63531]: DEBUG oslo_vmware.api [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52c4c044-9be0-945b-96bb-4cfc7ef17d0c, 'name': SearchDatastore_Task, 'duration_secs': 0.01001} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.369590] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6fba73c2-726f-421d-a7bf-82607ae6838d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.382195] env[63531]: ERROR root [req-12df086a-aaee-4fc0-9c79-a53119d48317 req-8bea5b09-c55f-421c-be42-47aed0180cfb service nova] Original exception being dropped: ['Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 377, in request_handler\n response = request(managed_object, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 586, in __call__\n return client.invoke(args, kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 728, in invoke\n result = self.send(soapenv, timeout=timeout)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 777, in send\n return self.process_reply(reply.message, None, None)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 840, in process_reply\n raise WebFault(fault, replyroot)\n', "suds.WebFault: Server raised fault: 'The object 'vim.VirtualMachine:vm-244774' has already been deleted or has not been completely created'\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 301, in _invoke_api\n return api_method(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 480, in get_object_property\n props = get_object_properties(vim, moref, [property_name],\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 360, in get_object_properties\n retrieve_result = vim.RetrievePropertiesEx(\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 413, in request_handler\n raise exceptions.VimFaultException(fault_list, fault_string,\n', "oslo_vmware.exceptions.VimFaultException: The object 'vim.VirtualMachine:vm-244774' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-244774' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-244774'}\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/session.py", line 123, in _call_method\n return self.invoke_api(module, method, self.vim, *args,\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 358, in invoke_api\n return _invoke_api(module, method, *args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 122, in func\n return evt.wait()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait\n result = hub.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch\n return self.greenlet.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/common/loopingcall.py", line 122, in _inner\n idle = self.f(*self.args, **self.kw)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 96, in _func\n result = f(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 341, in _invoke_api\n raise clazz(str(excep),\n', "oslo_vmware.exceptions.ManagedObjectNotFoundException: The object 'vim.VirtualMachine:vm-244774' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-244774' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-244774'}\n"]: nova.exception.InstanceNotFound: Instance e742d3af-5a45-4274-adfb-c6138bf84d6a could not be found. [ 1081.382459] env[63531]: DEBUG oslo_concurrency.lockutils [req-12df086a-aaee-4fc0-9c79-a53119d48317 req-8bea5b09-c55f-421c-be42-47aed0180cfb service nova] Releasing lock "e742d3af-5a45-4274-adfb-c6138bf84d6a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1081.382710] env[63531]: DEBUG nova.compute.manager [req-12df086a-aaee-4fc0-9c79-a53119d48317 req-8bea5b09-c55f-421c-be42-47aed0180cfb service nova] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Detach interface failed, port_id=5f33c6bf-e31f-422e-8611-c4efb3719681, reason: Instance e742d3af-5a45-4274-adfb-c6138bf84d6a could not be found. {{(pid=63531) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1081.389766] env[63531]: DEBUG oslo_vmware.api [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 1081.389766] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52d8ad1d-884e-13ec-92bc-ac9d239246fa" [ 1081.389766] env[63531]: _type = "Task" [ 1081.389766] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.399042] env[63531]: DEBUG oslo_vmware.api [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52d8ad1d-884e-13ec-92bc-ac9d239246fa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.402745] env[63531]: DEBUG nova.network.neutron [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Successfully updated port: 99a45e7c-1ddb-4c35-ac69-b2c33f976019 {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1081.485848] env[63531]: DEBUG oslo_concurrency.lockutils [req-effa0739-b90b-4f14-a1d9-69af34ae5984 req-f87d389d-1f95-4684-b573-d63f775770ce service nova] Releasing lock "refresh_cache-b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1081.519920] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31b7deaf-073d-47e9-9e68-c1df4e9d9646 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.531052] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] Skipping network cache update for instance because it is Building. {{(pid=63531) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 1081.531215] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Skipping network cache update for instance because it is Building. {{(pid=63531) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 1081.534062] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2a93866-3fb5-412d-b114-fc4df28bb6b9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.566686] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Acquiring lock "refresh_cache-8dec0cac-defc-42f2-ab56-6b3ae60ad858" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1081.566836] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Acquired lock "refresh_cache-8dec0cac-defc-42f2-ab56-6b3ae60ad858" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1081.567014] env[63531]: DEBUG nova.network.neutron [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Forcefully refreshing network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1081.567195] env[63531]: DEBUG nova.objects.instance [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Lazy-loading 'info_cache' on Instance uuid 8dec0cac-defc-42f2-ab56-6b3ae60ad858 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1081.572465] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25d09674-dacc-44bc-a11a-513423437ddd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.581528] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef24d900-1ac9-43b2-aef9-8ffa1bfa41b7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.599873] env[63531]: DEBUG nova.compute.provider_tree [None req-f2c55fcf-25ae-49ea-805d-5d27839b6186 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1081.858396] env[63531]: DEBUG nova.compute.manager [req-5502d5a4-932e-46c1-85b8-3a3ec865b22e req-b5fba64a-8ac0-4e5c-822a-4118f3688882 service nova] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Received event network-vif-plugged-99a45e7c-1ddb-4c35-ac69-b2c33f976019 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1081.858587] env[63531]: DEBUG oslo_concurrency.lockutils [req-5502d5a4-932e-46c1-85b8-3a3ec865b22e req-b5fba64a-8ac0-4e5c-822a-4118f3688882 service nova] Acquiring lock "49967454-a4f9-47f6-a2a9-0cd0c29e18d8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1081.858801] env[63531]: DEBUG oslo_concurrency.lockutils [req-5502d5a4-932e-46c1-85b8-3a3ec865b22e req-b5fba64a-8ac0-4e5c-822a-4118f3688882 service nova] Lock "49967454-a4f9-47f6-a2a9-0cd0c29e18d8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1081.858974] env[63531]: DEBUG oslo_concurrency.lockutils [req-5502d5a4-932e-46c1-85b8-3a3ec865b22e req-b5fba64a-8ac0-4e5c-822a-4118f3688882 service nova] Lock "49967454-a4f9-47f6-a2a9-0cd0c29e18d8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1081.859166] env[63531]: DEBUG nova.compute.manager [req-5502d5a4-932e-46c1-85b8-3a3ec865b22e req-b5fba64a-8ac0-4e5c-822a-4118f3688882 service nova] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] No waiting events found dispatching network-vif-plugged-99a45e7c-1ddb-4c35-ac69-b2c33f976019 {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1081.859375] env[63531]: WARNING nova.compute.manager [req-5502d5a4-932e-46c1-85b8-3a3ec865b22e req-b5fba64a-8ac0-4e5c-822a-4118f3688882 service nova] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Received unexpected event network-vif-plugged-99a45e7c-1ddb-4c35-ac69-b2c33f976019 for instance with vm_state building and task_state block_device_mapping. [ 1081.859546] env[63531]: DEBUG nova.compute.manager [req-5502d5a4-932e-46c1-85b8-3a3ec865b22e req-b5fba64a-8ac0-4e5c-822a-4118f3688882 service nova] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Received event network-changed-99a45e7c-1ddb-4c35-ac69-b2c33f976019 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1081.859703] env[63531]: DEBUG nova.compute.manager [req-5502d5a4-932e-46c1-85b8-3a3ec865b22e req-b5fba64a-8ac0-4e5c-822a-4118f3688882 service nova] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Refreshing instance network info cache due to event network-changed-99a45e7c-1ddb-4c35-ac69-b2c33f976019. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1081.859891] env[63531]: DEBUG oslo_concurrency.lockutils [req-5502d5a4-932e-46c1-85b8-3a3ec865b22e req-b5fba64a-8ac0-4e5c-822a-4118f3688882 service nova] Acquiring lock "refresh_cache-49967454-a4f9-47f6-a2a9-0cd0c29e18d8" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1081.860036] env[63531]: DEBUG oslo_concurrency.lockutils [req-5502d5a4-932e-46c1-85b8-3a3ec865b22e req-b5fba64a-8ac0-4e5c-822a-4118f3688882 service nova] Acquired lock "refresh_cache-49967454-a4f9-47f6-a2a9-0cd0c29e18d8" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1081.860200] env[63531]: DEBUG nova.network.neutron [req-5502d5a4-932e-46c1-85b8-3a3ec865b22e req-b5fba64a-8ac0-4e5c-822a-4118f3688882 service nova] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Refreshing network info cache for port 99a45e7c-1ddb-4c35-ac69-b2c33f976019 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1081.903605] env[63531]: DEBUG oslo_vmware.api [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52d8ad1d-884e-13ec-92bc-ac9d239246fa, 'name': SearchDatastore_Task, 'duration_secs': 0.010263} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.903892] env[63531]: DEBUG oslo_concurrency.lockutils [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1081.904195] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16/b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1081.904784] env[63531]: DEBUG oslo_concurrency.lockutils [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquiring lock "refresh_cache-49967454-a4f9-47f6-a2a9-0cd0c29e18d8" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1081.904908] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-41fecb43-9d75-451e-b140-84c67a1eb246 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.914032] env[63531]: DEBUG oslo_vmware.api [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 1081.914032] env[63531]: value = "task-1118596" [ 1081.914032] env[63531]: _type = "Task" [ 1081.914032] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.937247] env[63531]: DEBUG oslo_vmware.api [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118596, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.964294] env[63531]: DEBUG oslo_concurrency.lockutils [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "96c2910b-1e25-4053-9ad9-c155402ef0ac-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1081.964420] env[63531]: DEBUG oslo_concurrency.lockutils [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "96c2910b-1e25-4053-9ad9-c155402ef0ac-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1081.964522] env[63531]: DEBUG oslo_concurrency.lockutils [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "96c2910b-1e25-4053-9ad9-c155402ef0ac-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1082.104066] env[63531]: DEBUG nova.scheduler.client.report [None req-f2c55fcf-25ae-49ea-805d-5d27839b6186 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1082.411180] env[63531]: DEBUG nova.network.neutron [req-5502d5a4-932e-46c1-85b8-3a3ec865b22e req-b5fba64a-8ac0-4e5c-822a-4118f3688882 service nova] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1082.425106] env[63531]: DEBUG oslo_vmware.api [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118596, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.508768] env[63531]: DEBUG nova.network.neutron [req-5502d5a4-932e-46c1-85b8-3a3ec865b22e req-b5fba64a-8ac0-4e5c-822a-4118f3688882 service nova] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1082.613074] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f2c55fcf-25ae-49ea-805d-5d27839b6186 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.838s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1082.616252] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.642s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1082.618986] env[63531]: INFO nova.compute.claims [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1082.639496] env[63531]: INFO nova.scheduler.client.report [None req-f2c55fcf-25ae-49ea-805d-5d27839b6186 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Deleted allocations for instance 85058146-ba4c-4e9a-bbe7-048ed56167ab [ 1082.765618] env[63531]: DEBUG nova.compute.manager [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1082.766196] env[63531]: DEBUG nova.virt.hardware [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1082.766425] env[63531]: DEBUG nova.virt.hardware [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1082.766590] env[63531]: DEBUG nova.virt.hardware [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1082.766780] env[63531]: DEBUG nova.virt.hardware [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1082.766934] env[63531]: DEBUG nova.virt.hardware [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1082.767102] env[63531]: DEBUG nova.virt.hardware [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1082.767361] env[63531]: DEBUG nova.virt.hardware [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1082.767586] env[63531]: DEBUG nova.virt.hardware [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1082.767786] env[63531]: DEBUG nova.virt.hardware [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1082.768036] env[63531]: DEBUG nova.virt.hardware [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1082.768413] env[63531]: DEBUG nova.virt.hardware [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1082.769413] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a4ca239-94d7-462b-a733-f79dba2ae2f3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.778437] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d23a7d60-e7c4-48f1-a6c6-5c07717c84e0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.936590] env[63531]: DEBUG oslo_vmware.api [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118596, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.532026} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.936874] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16/b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1082.937120] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1082.937484] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0eeafd3f-0300-4e04-beb9-d6ec4eb3962c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.945269] env[63531]: DEBUG oslo_vmware.api [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 1082.945269] env[63531]: value = "task-1118597" [ 1082.945269] env[63531]: _type = "Task" [ 1082.945269] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.953624] env[63531]: DEBUG oslo_vmware.api [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118597, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.002405] env[63531]: DEBUG oslo_concurrency.lockutils [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "refresh_cache-96c2910b-1e25-4053-9ad9-c155402ef0ac" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1083.002687] env[63531]: DEBUG oslo_concurrency.lockutils [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquired lock "refresh_cache-96c2910b-1e25-4053-9ad9-c155402ef0ac" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.002924] env[63531]: DEBUG nova.network.neutron [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1083.012220] env[63531]: DEBUG oslo_concurrency.lockutils [req-5502d5a4-932e-46c1-85b8-3a3ec865b22e req-b5fba64a-8ac0-4e5c-822a-4118f3688882 service nova] Releasing lock "refresh_cache-49967454-a4f9-47f6-a2a9-0cd0c29e18d8" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1083.013815] env[63531]: DEBUG oslo_concurrency.lockutils [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquired lock "refresh_cache-49967454-a4f9-47f6-a2a9-0cd0c29e18d8" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.014094] env[63531]: DEBUG nova.network.neutron [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1083.147663] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f2c55fcf-25ae-49ea-805d-5d27839b6186 tempest-ServerDiskConfigTestJSON-1190369323 tempest-ServerDiskConfigTestJSON-1190369323-project-member] Lock "85058146-ba4c-4e9a-bbe7-048ed56167ab" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.060s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1083.351920] env[63531]: DEBUG nova.network.neutron [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Updating instance_info_cache with network_info: [{"id": "76060df9-1ff6-4c64-86ef-a8e0d4081cfd", "address": "fa:16:3e:52:f5:96", "network": {"id": "eac8786d-bc67-4916-8483-b53d9a25efb8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1604989690-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.241", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "abdae960d43b47bc95ab41f265d2c85b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db068f71-08cc-42d4-8ab6-17134c1585e5", "external-id": "nsx-vlan-transportzone-721", "segmentation_id": 721, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap76060df9-1f", "ovs_interfaceid": "76060df9-1ff6-4c64-86ef-a8e0d4081cfd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1083.455990] env[63531]: DEBUG oslo_vmware.api [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118597, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.17561} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.456342] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1083.457147] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ea8fd8c-7d65-4529-b8a8-1abd6f74ed3e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.481929] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16/b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1083.482223] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3d77323b-4f0a-4418-9d93-7f96bb2a0f20 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.504330] env[63531]: DEBUG oslo_vmware.api [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 1083.504330] env[63531]: value = "task-1118598" [ 1083.504330] env[63531]: _type = "Task" [ 1083.504330] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.516273] env[63531]: DEBUG oslo_vmware.api [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118598, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.558921] env[63531]: DEBUG nova.network.neutron [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1083.797685] env[63531]: DEBUG nova.network.neutron [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Updating instance_info_cache with network_info: [{"id": "99a45e7c-1ddb-4c35-ac69-b2c33f976019", "address": "fa:16:3e:c8:19:0d", "network": {"id": "eac8786d-bc67-4916-8483-b53d9a25efb8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1604989690-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "abdae960d43b47bc95ab41f265d2c85b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db068f71-08cc-42d4-8ab6-17134c1585e5", "external-id": "nsx-vlan-transportzone-721", "segmentation_id": 721, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap99a45e7c-1d", "ovs_interfaceid": "99a45e7c-1ddb-4c35-ac69-b2c33f976019", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1083.809745] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-847ab347-165a-4cff-ae07-2d728332b31b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.819978] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a5c6736-31db-4859-8dec-567954627934 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.857722] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Releasing lock "refresh_cache-8dec0cac-defc-42f2-ab56-6b3ae60ad858" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1083.857722] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Updated the network info_cache for instance {{(pid=63531) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1083.858480] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae227317-2433-42d6-8aa0-28c7d2088433 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.862728] env[63531]: DEBUG nova.network.neutron [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Updating instance_info_cache with network_info: [{"id": "b731766f-738c-4d77-a00e-98ed6c683e55", "address": "fa:16:3e:d5:26:02", "network": {"id": "0667c176-cfdf-4e43-b152-e00bdc224448", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-243715985-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.217", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "308eabafea9c4188a58a0f1c22074d2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8f580e6-1d86-41ee-9ebe-c531cb9299c6", "external-id": "nsx-vlan-transportzone-150", "segmentation_id": 150, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb731766f-73", "ovs_interfaceid": "b731766f-738c-4d77-a00e-98ed6c683e55", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1083.864387] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1083.868078] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1083.868078] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1083.868078] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1083.868078] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1083.868078] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1083.868078] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63531) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1083.868078] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager.update_available_resource {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1083.872060] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-984ac4c2-4fd4-48b2-a688-f9107e3cc35f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.885694] env[63531]: DEBUG nova.compute.provider_tree [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1084.022070] env[63531]: DEBUG oslo_vmware.api [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118598, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.216192] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "ccc8f366-5849-45a6-9745-ff9b2be9c3bb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1084.216626] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "ccc8f366-5849-45a6-9745-ff9b2be9c3bb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1084.301384] env[63531]: DEBUG oslo_concurrency.lockutils [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Releasing lock "refresh_cache-49967454-a4f9-47f6-a2a9-0cd0c29e18d8" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1084.301384] env[63531]: DEBUG nova.compute.manager [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Instance network_info: |[{"id": "99a45e7c-1ddb-4c35-ac69-b2c33f976019", "address": "fa:16:3e:c8:19:0d", "network": {"id": "eac8786d-bc67-4916-8483-b53d9a25efb8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1604989690-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "abdae960d43b47bc95ab41f265d2c85b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db068f71-08cc-42d4-8ab6-17134c1585e5", "external-id": "nsx-vlan-transportzone-721", "segmentation_id": 721, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap99a45e7c-1d", "ovs_interfaceid": "99a45e7c-1ddb-4c35-ac69-b2c33f976019", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1084.301384] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c8:19:0d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'db068f71-08cc-42d4-8ab6-17134c1585e5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '99a45e7c-1ddb-4c35-ac69-b2c33f976019', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1084.308698] env[63531]: DEBUG oslo.service.loopingcall [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1084.308923] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1084.309173] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2eec8dd5-234d-4b9c-8d3a-b276f2813749 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.329914] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1084.329914] env[63531]: value = "task-1118599" [ 1084.329914] env[63531]: _type = "Task" [ 1084.329914] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.338374] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118599, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.368738] env[63531]: DEBUG oslo_concurrency.lockutils [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Releasing lock "refresh_cache-96c2910b-1e25-4053-9ad9-c155402ef0ac" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1084.373191] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1084.388952] env[63531]: DEBUG nova.scheduler.client.report [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1084.515891] env[63531]: DEBUG oslo_vmware.api [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118598, 'name': ReconfigVM_Task, 'duration_secs': 0.865661} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.516285] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] Reconfigured VM instance instance-0000006a to attach disk [datastore1] b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16/b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1084.517016] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e3c4a4ef-6bdf-4ad7-9157-7b2f01e64584 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.525216] env[63531]: DEBUG oslo_vmware.api [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 1084.525216] env[63531]: value = "task-1118600" [ 1084.525216] env[63531]: _type = "Task" [ 1084.525216] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.537892] env[63531]: DEBUG oslo_vmware.api [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118600, 'name': Rename_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.720871] env[63531]: DEBUG nova.compute.manager [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1084.840690] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118599, 'name': CreateVM_Task, 'duration_secs': 0.331599} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.840879] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1084.841594] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'mount_device': '/dev/sda', 'attachment_id': '4db59766-5c8c-4c09-bb08-c72e703547ba', 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244800', 'volume_id': '422ea428-1dc4-40f2-8151-a5e18a596f9b', 'name': 'volume-422ea428-1dc4-40f2-8151-a5e18a596f9b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '49967454-a4f9-47f6-a2a9-0cd0c29e18d8', 'attached_at': '', 'detached_at': '', 'volume_id': '422ea428-1dc4-40f2-8151-a5e18a596f9b', 'serial': '422ea428-1dc4-40f2-8151-a5e18a596f9b'}, 'disk_bus': None, 'guest_format': None, 'delete_on_termination': True, 'boot_index': 0, 'device_type': None, 'volume_type': None}], 'swap': None} {{(pid=63531) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1084.841820] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Root volume attach. Driver type: vmdk {{(pid=63531) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 1084.842610] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5da7fa9e-fb96-407c-8b72-5fdbe8ef0d3e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.850532] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91d6343a-bd1e-41cd-8cdb-3ddec84fa23e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.856257] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5950152c-e6a0-4952-96c9-ddf586786d8c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.862752] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-15cd76c8-99f5-4a26-8acf-aee76aa291b2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.873986] env[63531]: DEBUG oslo_vmware.api [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 1084.873986] env[63531]: value = "task-1118601" [ 1084.873986] env[63531]: _type = "Task" [ 1084.873986] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.888387] env[63531]: DEBUG oslo_vmware.api [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118601, 'name': RelocateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.895016] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.278s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1084.895016] env[63531]: DEBUG nova.compute.manager [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1084.898232] env[63531]: DEBUG oslo_concurrency.lockutils [None req-cf55f1a9-a326-424b-abfc-9e7077a42e78 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.910s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1084.899448] env[63531]: DEBUG nova.objects.instance [None req-cf55f1a9-a326-424b-abfc-9e7077a42e78 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lazy-loading 'resources' on Instance uuid b7b2b7ac-2321-4cf3-b2ed-4877b893af63 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1084.900094] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daeb6da5-7117-47f4-b650-0da40ff823d9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.923791] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-399d6beb-8b71-4b3f-99b5-e6e6ba06b2ef {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.936917] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Updating instance '96c2910b-1e25-4053-9ad9-c155402ef0ac' progress to 83 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1085.036338] env[63531]: DEBUG oslo_vmware.api [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118600, 'name': Rename_Task, 'duration_secs': 0.166918} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.037022] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1085.037022] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-70f40af0-9d8c-494c-bf3e-8e3e04b3014c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.044039] env[63531]: DEBUG oslo_vmware.api [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 1085.044039] env[63531]: value = "task-1118602" [ 1085.044039] env[63531]: _type = "Task" [ 1085.044039] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.052293] env[63531]: DEBUG oslo_vmware.api [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118602, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.252496] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1085.390215] env[63531]: DEBUG oslo_vmware.api [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118601, 'name': RelocateVM_Task} progress is 42%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.412096] env[63531]: DEBUG nova.compute.utils [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1085.416018] env[63531]: DEBUG nova.compute.manager [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1085.416018] env[63531]: DEBUG nova.network.neutron [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1085.445743] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1085.446465] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-91665461-f3c6-4dd4-a61d-a6ecc17818b9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.458731] env[63531]: DEBUG oslo_vmware.api [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 1085.458731] env[63531]: value = "task-1118603" [ 1085.458731] env[63531]: _type = "Task" [ 1085.458731] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.470909] env[63531]: DEBUG oslo_vmware.api [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118603, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.474946] env[63531]: DEBUG nova.policy [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0372c22d1e9e4f66a0417643dab4a2ec', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '94a95667c00947ea9183a6307c569c90', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 1085.557105] env[63531]: DEBUG oslo_vmware.api [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118602, 'name': PowerOnVM_Task} progress is 78%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.627710] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be2be117-5aba-44fc-a40e-941b91513022 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.636711] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1c987ec-725f-4ef0-a43b-be116eab0205 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.682746] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-377cda56-b480-4855-95c6-7753eab50aeb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.694769] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2882678f-6fa2-44fd-8d93-214fdb98e891 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.721017] env[63531]: DEBUG nova.compute.provider_tree [None req-cf55f1a9-a326-424b-abfc-9e7077a42e78 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1085.768544] env[63531]: DEBUG nova.network.neutron [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] Successfully created port: 1765d3a3-d6ed-4150-aedf-45d90978133b {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1085.889616] env[63531]: DEBUG oslo_vmware.api [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118601, 'name': RelocateVM_Task} progress is 56%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.916539] env[63531]: DEBUG nova.compute.manager [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1085.970634] env[63531]: DEBUG oslo_vmware.api [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118603, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.058068] env[63531]: DEBUG oslo_vmware.api [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118602, 'name': PowerOnVM_Task} progress is 78%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.226055] env[63531]: DEBUG nova.scheduler.client.report [None req-cf55f1a9-a326-424b-abfc-9e7077a42e78 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1086.388814] env[63531]: DEBUG oslo_vmware.api [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118601, 'name': RelocateVM_Task} progress is 69%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.477272] env[63531]: DEBUG oslo_vmware.api [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118603, 'name': PowerOnVM_Task, 'duration_secs': 0.517437} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.477272] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1086.477272] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-592f7f10-0fee-489d-81c0-790438d86aaa tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Updating instance '96c2910b-1e25-4053-9ad9-c155402ef0ac' progress to 100 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1086.557972] env[63531]: DEBUG oslo_vmware.api [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118602, 'name': PowerOnVM_Task, 'duration_secs': 1.424397} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.558731] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1086.559139] env[63531]: INFO nova.compute.manager [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] Took 7.80 seconds to spawn the instance on the hypervisor. [ 1086.559560] env[63531]: DEBUG nova.compute.manager [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1086.561118] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b60ce315-bf77-436e-881f-c5f704f76a14 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.733932] env[63531]: DEBUG oslo_concurrency.lockutils [None req-cf55f1a9-a326-424b-abfc-9e7077a42e78 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.834s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1086.736620] env[63531]: DEBUG oslo_concurrency.lockutils [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 7.863s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1086.780218] env[63531]: INFO nova.scheduler.client.report [None req-cf55f1a9-a326-424b-abfc-9e7077a42e78 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Deleted allocations for instance b7b2b7ac-2321-4cf3-b2ed-4877b893af63 [ 1086.890708] env[63531]: DEBUG oslo_vmware.api [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118601, 'name': RelocateVM_Task} progress is 82%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.926718] env[63531]: DEBUG nova.compute.manager [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1086.956024] env[63531]: DEBUG nova.virt.hardware [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1086.956024] env[63531]: DEBUG nova.virt.hardware [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1086.956024] env[63531]: DEBUG nova.virt.hardware [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1086.956024] env[63531]: DEBUG nova.virt.hardware [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1086.956024] env[63531]: DEBUG nova.virt.hardware [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1086.956024] env[63531]: DEBUG nova.virt.hardware [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1086.956024] env[63531]: DEBUG nova.virt.hardware [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1086.956591] env[63531]: DEBUG nova.virt.hardware [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1086.956905] env[63531]: DEBUG nova.virt.hardware [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1086.957236] env[63531]: DEBUG nova.virt.hardware [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1086.957559] env[63531]: DEBUG nova.virt.hardware [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1086.958559] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eaa45df-8344-4b3e-bdb5-cce131b20ccb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.967421] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6ba87e1-d546-4875-8cd7-c4671da16b1d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.085364] env[63531]: INFO nova.compute.manager [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] Took 14.12 seconds to build instance. [ 1087.245139] env[63531]: DEBUG nova.objects.instance [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lazy-loading 'migration_context' on Instance uuid 18057005-1ff1-4e3d-bccc-8c89755d6db7 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1087.299235] env[63531]: DEBUG oslo_concurrency.lockutils [None req-cf55f1a9-a326-424b-abfc-9e7077a42e78 tempest-AttachVolumeTestJSON-1781071347 tempest-AttachVolumeTestJSON-1781071347-project-member] Lock "b7b2b7ac-2321-4cf3-b2ed-4877b893af63" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.504s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1087.393372] env[63531]: DEBUG oslo_vmware.api [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118601, 'name': RelocateVM_Task} progress is 97%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.564422] env[63531]: DEBUG nova.compute.manager [req-6d2c9e94-5dd3-4329-be4b-cb123e630655 req-9929b979-6cbb-4fa7-9e73-68ac7f9ce931 service nova] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] Received event network-vif-plugged-1765d3a3-d6ed-4150-aedf-45d90978133b {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1087.564422] env[63531]: DEBUG oslo_concurrency.lockutils [req-6d2c9e94-5dd3-4329-be4b-cb123e630655 req-9929b979-6cbb-4fa7-9e73-68ac7f9ce931 service nova] Acquiring lock "148c6ecd-354e-4076-b52b-c13a2119da7e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1087.564422] env[63531]: DEBUG oslo_concurrency.lockutils [req-6d2c9e94-5dd3-4329-be4b-cb123e630655 req-9929b979-6cbb-4fa7-9e73-68ac7f9ce931 service nova] Lock "148c6ecd-354e-4076-b52b-c13a2119da7e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1087.565392] env[63531]: DEBUG oslo_concurrency.lockutils [req-6d2c9e94-5dd3-4329-be4b-cb123e630655 req-9929b979-6cbb-4fa7-9e73-68ac7f9ce931 service nova] Lock "148c6ecd-354e-4076-b52b-c13a2119da7e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1087.565795] env[63531]: DEBUG nova.compute.manager [req-6d2c9e94-5dd3-4329-be4b-cb123e630655 req-9929b979-6cbb-4fa7-9e73-68ac7f9ce931 service nova] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] No waiting events found dispatching network-vif-plugged-1765d3a3-d6ed-4150-aedf-45d90978133b {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1087.566121] env[63531]: WARNING nova.compute.manager [req-6d2c9e94-5dd3-4329-be4b-cb123e630655 req-9929b979-6cbb-4fa7-9e73-68ac7f9ce931 service nova] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] Received unexpected event network-vif-plugged-1765d3a3-d6ed-4150-aedf-45d90978133b for instance with vm_state building and task_state spawning. [ 1087.588109] env[63531]: DEBUG oslo_concurrency.lockutils [None req-64db11da-bf28-4807-b9b2-8d7b7366feca tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.646s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1087.771703] env[63531]: DEBUG nova.scheduler.client.report [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Refreshing inventories for resource provider 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1087.792023] env[63531]: DEBUG nova.scheduler.client.report [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Updating ProviderTree inventory for provider 9996a314-a62e-4282-8252-f387e7e3fabb from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1087.792023] env[63531]: DEBUG nova.compute.provider_tree [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Updating inventory in ProviderTree for provider 9996a314-a62e-4282-8252-f387e7e3fabb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1087.808034] env[63531]: DEBUG nova.scheduler.client.report [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Refreshing aggregate associations for resource provider 9996a314-a62e-4282-8252-f387e7e3fabb, aggregates: None {{(pid=63531) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1087.831099] env[63531]: DEBUG nova.scheduler.client.report [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Refreshing trait associations for resource provider 9996a314-a62e-4282-8252-f387e7e3fabb, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE {{(pid=63531) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1087.897286] env[63531]: DEBUG oslo_vmware.api [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118601, 'name': RelocateVM_Task} progress is 98%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.026446] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95e1de78-5a88-4ec1-a33f-4d8711e8793e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.036987] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3b7a519-f7bc-481e-aa6e-5f3475a8556c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.071939] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f116b45d-08e5-47c9-bc19-02afedf43aed {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.081983] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8382e80f-b102-4e8b-86dd-2b32d156e2d1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.098828] env[63531]: DEBUG nova.compute.provider_tree [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Updating inventory in ProviderTree for provider 9996a314-a62e-4282-8252-f387e7e3fabb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1088.125548] env[63531]: DEBUG nova.network.neutron [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] Successfully updated port: 1765d3a3-d6ed-4150-aedf-45d90978133b {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1088.169559] env[63531]: DEBUG nova.compute.manager [req-74abd423-1f80-4152-950f-f5c614678851 req-2d91b2a7-a1df-44cc-a8a9-58e5cfa84ed6 service nova] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] Received event network-changed-1765d3a3-d6ed-4150-aedf-45d90978133b {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1088.169769] env[63531]: DEBUG nova.compute.manager [req-74abd423-1f80-4152-950f-f5c614678851 req-2d91b2a7-a1df-44cc-a8a9-58e5cfa84ed6 service nova] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] Refreshing instance network info cache due to event network-changed-1765d3a3-d6ed-4150-aedf-45d90978133b. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1088.170007] env[63531]: DEBUG oslo_concurrency.lockutils [req-74abd423-1f80-4152-950f-f5c614678851 req-2d91b2a7-a1df-44cc-a8a9-58e5cfa84ed6 service nova] Acquiring lock "refresh_cache-148c6ecd-354e-4076-b52b-c13a2119da7e" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1088.170162] env[63531]: DEBUG oslo_concurrency.lockutils [req-74abd423-1f80-4152-950f-f5c614678851 req-2d91b2a7-a1df-44cc-a8a9-58e5cfa84ed6 service nova] Acquired lock "refresh_cache-148c6ecd-354e-4076-b52b-c13a2119da7e" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1088.170383] env[63531]: DEBUG nova.network.neutron [req-74abd423-1f80-4152-950f-f5c614678851 req-2d91b2a7-a1df-44cc-a8a9-58e5cfa84ed6 service nova] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] Refreshing network info cache for port 1765d3a3-d6ed-4150-aedf-45d90978133b {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1088.392553] env[63531]: DEBUG oslo_vmware.api [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118601, 'name': RelocateVM_Task, 'duration_secs': 3.417937} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.394179] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Volume attach. Driver type: vmdk {{(pid=63531) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1088.394521] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244800', 'volume_id': '422ea428-1dc4-40f2-8151-a5e18a596f9b', 'name': 'volume-422ea428-1dc4-40f2-8151-a5e18a596f9b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '49967454-a4f9-47f6-a2a9-0cd0c29e18d8', 'attached_at': '', 'detached_at': '', 'volume_id': '422ea428-1dc4-40f2-8151-a5e18a596f9b', 'serial': '422ea428-1dc4-40f2-8151-a5e18a596f9b'} {{(pid=63531) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1088.399485] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88f80bca-5044-492f-9274-08a322c6721f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.422168] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f70ca16-24f7-40bd-8f65-51ba46334918 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.457093] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Reconfiguring VM instance instance-0000006b to attach disk [datastore2] volume-422ea428-1dc4-40f2-8151-a5e18a596f9b/volume-422ea428-1dc4-40f2-8151-a5e18a596f9b.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1088.457093] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-93ddcf2b-dd93-4bf0-b806-8f328375837e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.479301] env[63531]: DEBUG oslo_vmware.api [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 1088.479301] env[63531]: value = "task-1118605" [ 1088.479301] env[63531]: _type = "Task" [ 1088.479301] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.489324] env[63531]: DEBUG oslo_vmware.api [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118605, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.628388] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "refresh_cache-148c6ecd-354e-4076-b52b-c13a2119da7e" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1088.637548] env[63531]: DEBUG nova.scheduler.client.report [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Updated inventory for provider 9996a314-a62e-4282-8252-f387e7e3fabb with generation 144 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1088.637865] env[63531]: DEBUG nova.compute.provider_tree [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Updating resource provider 9996a314-a62e-4282-8252-f387e7e3fabb generation from 144 to 145 during operation: update_inventory {{(pid=63531) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1088.638109] env[63531]: DEBUG nova.compute.provider_tree [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Updating inventory in ProviderTree for provider 9996a314-a62e-4282-8252-f387e7e3fabb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1088.730628] env[63531]: DEBUG nova.network.neutron [req-74abd423-1f80-4152-950f-f5c614678851 req-2d91b2a7-a1df-44cc-a8a9-58e5cfa84ed6 service nova] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1088.864680] env[63531]: DEBUG nova.network.neutron [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Port b731766f-738c-4d77-a00e-98ed6c683e55 binding to destination host cpu-1 is already ACTIVE {{(pid=63531) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1088.865012] env[63531]: DEBUG oslo_concurrency.lockutils [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "refresh_cache-96c2910b-1e25-4053-9ad9-c155402ef0ac" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1088.865174] env[63531]: DEBUG oslo_concurrency.lockutils [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquired lock "refresh_cache-96c2910b-1e25-4053-9ad9-c155402ef0ac" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1088.865345] env[63531]: DEBUG nova.network.neutron [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1088.902151] env[63531]: DEBUG nova.network.neutron [req-74abd423-1f80-4152-950f-f5c614678851 req-2d91b2a7-a1df-44cc-a8a9-58e5cfa84ed6 service nova] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1088.903223] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a13a31cb-b684-4c2f-ae54-622a7c633f10 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1088.903448] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a13a31cb-b684-4c2f-ae54-622a7c633f10 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1088.903654] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a13a31cb-b684-4c2f-ae54-622a7c633f10 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1088.903858] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a13a31cb-b684-4c2f-ae54-622a7c633f10 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1088.904472] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a13a31cb-b684-4c2f-ae54-622a7c633f10 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1088.906284] env[63531]: INFO nova.compute.manager [None req-a13a31cb-b684-4c2f-ae54-622a7c633f10 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] Terminating instance [ 1088.910696] env[63531]: DEBUG nova.compute.manager [None req-a13a31cb-b684-4c2f-ae54-622a7c633f10 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1088.910923] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a13a31cb-b684-4c2f-ae54-622a7c633f10 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1088.912861] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f1b5325-a487-4569-bbf0-656a644e6e6a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.935225] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-a13a31cb-b684-4c2f-ae54-622a7c633f10 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1088.935528] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e274aedb-4cca-4a73-8dca-2a3b9b1bde1f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.942959] env[63531]: DEBUG oslo_vmware.api [None req-a13a31cb-b684-4c2f-ae54-622a7c633f10 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 1088.942959] env[63531]: value = "task-1118606" [ 1088.942959] env[63531]: _type = "Task" [ 1088.942959] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.954750] env[63531]: DEBUG oslo_vmware.api [None req-a13a31cb-b684-4c2f-ae54-622a7c633f10 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118606, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.994564] env[63531]: DEBUG oslo_vmware.api [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118605, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.407425] env[63531]: DEBUG oslo_concurrency.lockutils [req-74abd423-1f80-4152-950f-f5c614678851 req-2d91b2a7-a1df-44cc-a8a9-58e5cfa84ed6 service nova] Releasing lock "refresh_cache-148c6ecd-354e-4076-b52b-c13a2119da7e" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1089.407859] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquired lock "refresh_cache-148c6ecd-354e-4076-b52b-c13a2119da7e" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1089.408626] env[63531]: DEBUG nova.network.neutron [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1089.464029] env[63531]: DEBUG oslo_vmware.api [None req-a13a31cb-b684-4c2f-ae54-622a7c633f10 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118606, 'name': PowerOffVM_Task, 'duration_secs': 0.302536} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.464029] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-a13a31cb-b684-4c2f-ae54-622a7c633f10 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1089.464197] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a13a31cb-b684-4c2f-ae54-622a7c633f10 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1089.465063] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6dbaf52d-aef0-479e-abd1-d955317692d6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.500158] env[63531]: DEBUG oslo_vmware.api [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118605, 'name': ReconfigVM_Task, 'duration_secs': 0.786042} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.500158] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Reconfigured VM instance instance-0000006b to attach disk [datastore2] volume-422ea428-1dc4-40f2-8151-a5e18a596f9b/volume-422ea428-1dc4-40f2-8151-a5e18a596f9b.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1089.503615] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6a58d2e2-3cc9-476d-bea4-5537555eb486 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.519739] env[63531]: DEBUG oslo_vmware.api [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 1089.519739] env[63531]: value = "task-1118608" [ 1089.519739] env[63531]: _type = "Task" [ 1089.519739] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.532813] env[63531]: DEBUG oslo_vmware.api [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118608, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.565925] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a13a31cb-b684-4c2f-ae54-622a7c633f10 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1089.566178] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a13a31cb-b684-4c2f-ae54-622a7c633f10 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1089.568073] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-a13a31cb-b684-4c2f-ae54-622a7c633f10 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Deleting the datastore file [datastore1] b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1089.568073] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ca3dd71e-a51c-4635-a44a-38151ba06a9c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.577134] env[63531]: DEBUG oslo_vmware.api [None req-a13a31cb-b684-4c2f-ae54-622a7c633f10 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 1089.577134] env[63531]: value = "task-1118609" [ 1089.577134] env[63531]: _type = "Task" [ 1089.577134] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.587871] env[63531]: DEBUG oslo_vmware.api [None req-a13a31cb-b684-4c2f-ae54-622a7c633f10 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118609, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.640617] env[63531]: DEBUG nova.network.neutron [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Updating instance_info_cache with network_info: [{"id": "b731766f-738c-4d77-a00e-98ed6c683e55", "address": "fa:16:3e:d5:26:02", "network": {"id": "0667c176-cfdf-4e43-b152-e00bdc224448", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-243715985-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.217", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "308eabafea9c4188a58a0f1c22074d2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8f580e6-1d86-41ee-9ebe-c531cb9299c6", "external-id": "nsx-vlan-transportzone-150", "segmentation_id": 150, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb731766f-73", "ovs_interfaceid": "b731766f-738c-4d77-a00e-98ed6c683e55", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1089.649861] env[63531]: DEBUG oslo_concurrency.lockutils [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.913s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1089.656521] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 5.283s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1089.656643] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1089.656690] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63531) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1089.657044] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.405s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1089.658741] env[63531]: INFO nova.compute.claims [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1089.663763] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03fcd4ca-293d-4477-84c5-aee198b51b38 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.687602] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e2ed96a-8370-4f93-9b54-abe683787d9d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.703529] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfd09b5d-0ee1-408a-9b08-5fea9fa2a856 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.711650] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50ac11d3-f230-47a8-b00a-9b2214382f92 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.745253] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179622MB free_disk=170GB free_vcpus=48 pci_devices=None {{(pid=63531) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1089.745253] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1089.944029] env[63531]: DEBUG nova.network.neutron [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1090.035371] env[63531]: DEBUG oslo_vmware.api [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118608, 'name': ReconfigVM_Task, 'duration_secs': 0.143419} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.035714] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244800', 'volume_id': '422ea428-1dc4-40f2-8151-a5e18a596f9b', 'name': 'volume-422ea428-1dc4-40f2-8151-a5e18a596f9b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '49967454-a4f9-47f6-a2a9-0cd0c29e18d8', 'attached_at': '', 'detached_at': '', 'volume_id': '422ea428-1dc4-40f2-8151-a5e18a596f9b', 'serial': '422ea428-1dc4-40f2-8151-a5e18a596f9b'} {{(pid=63531) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1090.037023] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2624110a-fa98-4b74-8ffe-8ef1a6e5cbbe {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.045168] env[63531]: DEBUG oslo_vmware.api [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 1090.045168] env[63531]: value = "task-1118610" [ 1090.045168] env[63531]: _type = "Task" [ 1090.045168] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.058326] env[63531]: DEBUG oslo_vmware.api [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118610, 'name': Rename_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.091457] env[63531]: DEBUG oslo_vmware.api [None req-a13a31cb-b684-4c2f-ae54-622a7c633f10 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118609, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.456507} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.091829] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-a13a31cb-b684-4c2f-ae54-622a7c633f10 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1090.092111] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a13a31cb-b684-4c2f-ae54-622a7c633f10 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1090.092383] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a13a31cb-b684-4c2f-ae54-622a7c633f10 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1090.092648] env[63531]: INFO nova.compute.manager [None req-a13a31cb-b684-4c2f-ae54-622a7c633f10 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] Took 1.18 seconds to destroy the instance on the hypervisor. [ 1090.093031] env[63531]: DEBUG oslo.service.loopingcall [None req-a13a31cb-b684-4c2f-ae54-622a7c633f10 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1090.093297] env[63531]: DEBUG nova.compute.manager [-] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1090.093431] env[63531]: DEBUG nova.network.neutron [-] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1090.108217] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Acquiring lock "fbf63fa0-c8ec-41c7-878e-1446537b7b49" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1090.108485] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Lock "fbf63fa0-c8ec-41c7-878e-1446537b7b49" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1090.144116] env[63531]: DEBUG oslo_concurrency.lockutils [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Releasing lock "refresh_cache-96c2910b-1e25-4053-9ad9-c155402ef0ac" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1090.161158] env[63531]: DEBUG nova.network.neutron [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] Updating instance_info_cache with network_info: [{"id": "1765d3a3-d6ed-4150-aedf-45d90978133b", "address": "fa:16:3e:2c:94:f4", "network": {"id": "ea2d47d3-a502-4afd-9bc5-ee4a37017901", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1361282648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94a95667c00947ea9183a6307c569c90", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6c45fd-e930-495a-9cb7-df84eda443b1", "external-id": "nsx-vlan-transportzone-407", "segmentation_id": 407, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1765d3a3-d6", "ovs_interfaceid": "1765d3a3-d6ed-4150-aedf-45d90978133b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1090.430616] env[63531]: DEBUG nova.compute.manager [req-6f60fbdd-ef4b-42e3-9943-f695e426605f req-a95878d6-c9fb-4b65-8d61-9ee2d2dd24f9 service nova] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] Received event network-vif-deleted-6ae9a4f8-494f-4ba5-997a-0dcf67e4320a {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1090.430845] env[63531]: INFO nova.compute.manager [req-6f60fbdd-ef4b-42e3-9943-f695e426605f req-a95878d6-c9fb-4b65-8d61-9ee2d2dd24f9 service nova] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] Neutron deleted interface 6ae9a4f8-494f-4ba5-997a-0dcf67e4320a; detaching it from the instance and deleting it from the info cache [ 1090.431035] env[63531]: DEBUG nova.network.neutron [req-6f60fbdd-ef4b-42e3-9943-f695e426605f req-a95878d6-c9fb-4b65-8d61-9ee2d2dd24f9 service nova] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1090.557670] env[63531]: DEBUG oslo_vmware.api [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118610, 'name': Rename_Task, 'duration_secs': 0.171342} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.560020] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1090.560020] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3d6f3968-f5df-4ef8-b351-54327c44ee1a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.568737] env[63531]: DEBUG oslo_vmware.api [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 1090.568737] env[63531]: value = "task-1118611" [ 1090.568737] env[63531]: _type = "Task" [ 1090.568737] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.579641] env[63531]: DEBUG oslo_vmware.api [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118611, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.611568] env[63531]: DEBUG nova.compute.manager [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] [instance: fbf63fa0-c8ec-41c7-878e-1446537b7b49] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1090.648417] env[63531]: DEBUG nova.compute.manager [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=63531) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 1090.648417] env[63531]: DEBUG oslo_concurrency.lockutils [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1090.667267] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Releasing lock "refresh_cache-148c6ecd-354e-4076-b52b-c13a2119da7e" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1090.668245] env[63531]: DEBUG nova.compute.manager [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] Instance network_info: |[{"id": "1765d3a3-d6ed-4150-aedf-45d90978133b", "address": "fa:16:3e:2c:94:f4", "network": {"id": "ea2d47d3-a502-4afd-9bc5-ee4a37017901", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1361282648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94a95667c00947ea9183a6307c569c90", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6c45fd-e930-495a-9cb7-df84eda443b1", "external-id": "nsx-vlan-transportzone-407", "segmentation_id": 407, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1765d3a3-d6", "ovs_interfaceid": "1765d3a3-d6ed-4150-aedf-45d90978133b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1090.668245] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2c:94:f4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2b6c45fd-e930-495a-9cb7-df84eda443b1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1765d3a3-d6ed-4150-aedf-45d90978133b', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1090.678173] env[63531]: DEBUG oslo.service.loopingcall [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1090.678783] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1090.679051] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a80bdf52-4c23-45cd-aafe-ccf2e597c479 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.712363] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1090.712363] env[63531]: value = "task-1118612" [ 1090.712363] env[63531]: _type = "Task" [ 1090.712363] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.728181] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118612, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.909732] env[63531]: DEBUG nova.network.neutron [-] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1090.919661] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2412083d-7278-42c0-a678-d7a4cd2dc6ad {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.929488] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9970f55f-e04f-4f6e-a58d-5905cb613ab1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.935062] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a3c5b6d2-7d34-441d-beba-8ce73d92857b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.964264] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70900f93-2df4-484d-ab5f-fc0c3066d2b2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.971528] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03df1918-34e8-4f04-9631-99a2230eed17 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.992443] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2012745-0109-4e23-aa5f-9cf9c6a31ed1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.012185] env[63531]: DEBUG nova.compute.provider_tree [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1091.025734] env[63531]: DEBUG nova.compute.manager [req-6f60fbdd-ef4b-42e3-9943-f695e426605f req-a95878d6-c9fb-4b65-8d61-9ee2d2dd24f9 service nova] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] Detach interface failed, port_id=6ae9a4f8-494f-4ba5-997a-0dcf67e4320a, reason: Instance b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16 could not be found. {{(pid=63531) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1091.027368] env[63531]: DEBUG nova.scheduler.client.report [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1091.080015] env[63531]: DEBUG oslo_vmware.api [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118611, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.141389] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.223451] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118612, 'name': CreateVM_Task, 'duration_secs': 0.42412} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.224521] env[63531]: INFO nova.compute.manager [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Swapping old allocation on dict_keys(['9996a314-a62e-4282-8252-f387e7e3fabb']) held by migration b6672930-332c-4293-8ab4-8ad450b70f9d for instance [ 1091.224695] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1091.227247] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1091.227454] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1091.227786] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1091.228071] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5ee77e6f-77e0-4fae-8818-6726b5ced3cb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.235545] env[63531]: DEBUG oslo_vmware.api [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 1091.235545] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]529d2580-4088-24dd-503b-e9e7f04e99e4" [ 1091.235545] env[63531]: _type = "Task" [ 1091.235545] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.249832] env[63531]: DEBUG oslo_vmware.api [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]529d2580-4088-24dd-503b-e9e7f04e99e4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.254261] env[63531]: DEBUG nova.scheduler.client.report [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Overwriting current allocation {'allocations': {'9996a314-a62e-4282-8252-f387e7e3fabb': {'resources': {'VCPU': 1, 'MEMORY_MB': 256, 'DISK_GB': 1}, 'generation': 146}}, 'project_id': '57e9bf4ee0f1463fa67b0f8deff30fdc', 'user_id': '2ec68c1860c5484cb8094370baa80c98', 'consumer_generation': 1} on consumer 18057005-1ff1-4e3d-bccc-8c89755d6db7 {{(pid=63531) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1091.340481] env[63531]: DEBUG oslo_concurrency.lockutils [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquiring lock "refresh_cache-18057005-1ff1-4e3d-bccc-8c89755d6db7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1091.340833] env[63531]: DEBUG oslo_concurrency.lockutils [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquired lock "refresh_cache-18057005-1ff1-4e3d-bccc-8c89755d6db7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1091.341155] env[63531]: DEBUG nova.network.neutron [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1091.412659] env[63531]: INFO nova.compute.manager [-] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] Took 1.32 seconds to deallocate network for instance. [ 1091.533485] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.876s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1091.533855] env[63531]: DEBUG nova.compute.manager [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1091.538126] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 1.793s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.581030] env[63531]: DEBUG oslo_vmware.api [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118611, 'name': PowerOnVM_Task, 'duration_secs': 0.562208} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.581338] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1091.581546] env[63531]: INFO nova.compute.manager [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Took 8.82 seconds to spawn the instance on the hypervisor. [ 1091.581736] env[63531]: DEBUG nova.compute.manager [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1091.582533] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8c186d0-6e1f-4f7b-ab86-35a56899b245 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.749036] env[63531]: DEBUG oslo_vmware.api [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]529d2580-4088-24dd-503b-e9e7f04e99e4, 'name': SearchDatastore_Task, 'duration_secs': 0.036518} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.749036] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1091.749486] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1091.749486] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1091.749486] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1091.749796] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1091.750105] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-65cbabc7-a6fa-4c7b-a9a1-fdd16c53c3d9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.760309] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1091.760518] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1091.761257] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0d112c91-a4d7-4b4c-8d42-8ec6de2e5b0c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.767164] env[63531]: DEBUG oslo_vmware.api [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 1091.767164] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52390133-23d3-1b6a-6cd4-f22f512c5a2d" [ 1091.767164] env[63531]: _type = "Task" [ 1091.767164] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.775277] env[63531]: DEBUG oslo_vmware.api [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52390133-23d3-1b6a-6cd4-f22f512c5a2d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.920771] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a13a31cb-b684-4c2f-ae54-622a7c633f10 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1092.042251] env[63531]: DEBUG nova.compute.utils [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1092.052294] env[63531]: DEBUG nova.compute.manager [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1092.052294] env[63531]: DEBUG nova.network.neutron [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1092.104950] env[63531]: INFO nova.compute.manager [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Took 19.06 seconds to build instance. [ 1092.114622] env[63531]: DEBUG nova.policy [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '36cb2f0de54f417384fa1fb2210de4e4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8ba8b284854242c392aec5326e996239', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 1092.117357] env[63531]: DEBUG nova.network.neutron [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Updating instance_info_cache with network_info: [{"id": "af2485f6-5ce2-4a83-9d95-7bea85cebff2", "address": "fa:16:3e:f0:7d:92", "network": {"id": "c4f4c994-36ab-42ab-94f8-78f29cbafa50", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-994507698-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "57e9bf4ee0f1463fa67b0f8deff30fdc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "767a3a48-41d4-4a0c-961d-0024837f63bd", "external-id": "nsx-vlan-transportzone-141", "segmentation_id": 141, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaf2485f6-5c", "ovs_interfaceid": "af2485f6-5ce2-4a83-9d95-7bea85cebff2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1092.284346] env[63531]: DEBUG oslo_vmware.api [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52390133-23d3-1b6a-6cd4-f22f512c5a2d, 'name': SearchDatastore_Task, 'duration_secs': 0.01053} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.285191] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2de7f22e-f290-4247-aba8-882f37b82d75 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.290988] env[63531]: DEBUG oslo_vmware.api [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 1092.290988] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]520b288b-01ab-7338-abfa-64f632d1227e" [ 1092.290988] env[63531]: _type = "Task" [ 1092.290988] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.299465] env[63531]: DEBUG oslo_vmware.api [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]520b288b-01ab-7338-abfa-64f632d1227e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.413767] env[63531]: DEBUG nova.network.neutron [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Successfully created port: e9692fb2-f3a9-455c-8317-b533a6af48f7 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1092.551315] env[63531]: DEBUG nova.compute.manager [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1092.554722] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Applying migration context for instance 96c2910b-1e25-4053-9ad9-c155402ef0ac as it has an incoming, in-progress migration 063e7061-74ea-430a-93f4-e2b436fafba7. Migration status is reverting {{(pid=63531) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1092.555928] env[63531]: INFO nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Updating resource usage from migration 063e7061-74ea-430a-93f4-e2b436fafba7 [ 1092.585821] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 8dec0cac-defc-42f2-ab56-6b3ae60ad858 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1092.585981] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1092.586120] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1092.586242] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 18057005-1ff1-4e3d-bccc-8c89755d6db7 actively managed on this compute host and has allocations in placement: {'resources': {'VCPU': 1, 'MEMORY_MB': 192, 'DISK_GB': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1092.586360] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 89f19669-53a9-4425-a3c6-d5f8873b707e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1092.586478] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Migration 063e7061-74ea-430a-93f4-e2b436fafba7 is active on this compute host and has allocations in placement: {'resources': {'VCPU': 1, 'MEMORY_MB': 192, 'DISK_GB': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1092.586798] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 96c2910b-1e25-4053-9ad9-c155402ef0ac actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1092.587067] env[63531]: WARNING nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1092.587125] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 49967454-a4f9-47f6-a2a9-0cd0c29e18d8 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1092.587252] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 148c6ecd-354e-4076-b52b-c13a2119da7e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1092.587373] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance ccc8f366-5849-45a6-9745-ff9b2be9c3bb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1092.607404] env[63531]: DEBUG oslo_concurrency.lockutils [None req-74aca6a5-62be-4ae3-9644-151faaac9a95 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "49967454-a4f9-47f6-a2a9-0cd0c29e18d8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.587s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1092.619568] env[63531]: DEBUG oslo_concurrency.lockutils [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Releasing lock "refresh_cache-18057005-1ff1-4e3d-bccc-8c89755d6db7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1092.620807] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84ea61f7-c926-4ef6-aba1-653519ace1dd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.629145] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ef6eb89-049f-4da7-a104-0702e5def789 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.804650] env[63531]: DEBUG oslo_vmware.api [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]520b288b-01ab-7338-abfa-64f632d1227e, 'name': SearchDatastore_Task, 'duration_secs': 0.013883} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.804945] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1092.805536] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] 148c6ecd-354e-4076-b52b-c13a2119da7e/148c6ecd-354e-4076-b52b-c13a2119da7e.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1092.805896] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-536a0404-8818-4d89-8a4b-258962d7768a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.814383] env[63531]: DEBUG oslo_vmware.api [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 1092.814383] env[63531]: value = "task-1118614" [ 1092.814383] env[63531]: _type = "Task" [ 1092.814383] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.823761] env[63531]: DEBUG oslo_vmware.api [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118614, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.090552] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance fbf63fa0-c8ec-41c7-878e-1446537b7b49 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1093.090808] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Total usable vcpus: 48, total allocated vcpus: 10 {{(pid=63531) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1093.091017] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2496MB phys_disk=200GB used_disk=9GB total_vcpus=48 used_vcpus=10 pci_stats=[] {{(pid=63531) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1093.275145] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dd548f3-1afe-4d0f-9784-5e314174dc51 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.285528] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19a7c86f-5a0b-4815-aa8e-bf5608502a5b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.323797] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-511bc990-2a3f-47e9-b296-fc0e37cc9bd6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.334268] env[63531]: DEBUG oslo_vmware.api [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118614, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.337835] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16d82c43-c0fa-4439-ac60-3017669e1f6a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.357239] env[63531]: DEBUG nova.compute.provider_tree [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1093.499840] env[63531]: DEBUG nova.compute.manager [req-0b98624c-4cdc-468e-a53d-8dcb137dc7be req-d940b3aa-34de-4f45-a4f6-8fc53a306f8a service nova] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Received event network-changed-76060df9-1ff6-4c64-86ef-a8e0d4081cfd {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1093.500053] env[63531]: DEBUG nova.compute.manager [req-0b98624c-4cdc-468e-a53d-8dcb137dc7be req-d940b3aa-34de-4f45-a4f6-8fc53a306f8a service nova] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Refreshing instance network info cache due to event network-changed-76060df9-1ff6-4c64-86ef-a8e0d4081cfd. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1093.500334] env[63531]: DEBUG oslo_concurrency.lockutils [req-0b98624c-4cdc-468e-a53d-8dcb137dc7be req-d940b3aa-34de-4f45-a4f6-8fc53a306f8a service nova] Acquiring lock "refresh_cache-8dec0cac-defc-42f2-ab56-6b3ae60ad858" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1093.500423] env[63531]: DEBUG oslo_concurrency.lockutils [req-0b98624c-4cdc-468e-a53d-8dcb137dc7be req-d940b3aa-34de-4f45-a4f6-8fc53a306f8a service nova] Acquired lock "refresh_cache-8dec0cac-defc-42f2-ab56-6b3ae60ad858" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1093.500593] env[63531]: DEBUG nova.network.neutron [req-0b98624c-4cdc-468e-a53d-8dcb137dc7be req-d940b3aa-34de-4f45-a4f6-8fc53a306f8a service nova] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Refreshing network info cache for port 76060df9-1ff6-4c64-86ef-a8e0d4081cfd {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1093.564730] env[63531]: DEBUG nova.compute.manager [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1093.598021] env[63531]: DEBUG nova.virt.hardware [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1093.598021] env[63531]: DEBUG nova.virt.hardware [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1093.598021] env[63531]: DEBUG nova.virt.hardware [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1093.598021] env[63531]: DEBUG nova.virt.hardware [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1093.598021] env[63531]: DEBUG nova.virt.hardware [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1093.598021] env[63531]: DEBUG nova.virt.hardware [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1093.598714] env[63531]: DEBUG nova.virt.hardware [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1093.599045] env[63531]: DEBUG nova.virt.hardware [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1093.599355] env[63531]: DEBUG nova.virt.hardware [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1093.599695] env[63531]: DEBUG nova.virt.hardware [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1093.600016] env[63531]: DEBUG nova.virt.hardware [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1093.601053] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1a8fe66-ac0e-4d2c-998c-23c8be873977 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.614049] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-775b38b4-07b0-4737-a36d-bb5a442daf23 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.742188] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1093.742714] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-af542f4c-33e9-4ad7-a3b4-507258f3a531 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.752140] env[63531]: DEBUG oslo_vmware.api [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 1093.752140] env[63531]: value = "task-1118615" [ 1093.752140] env[63531]: _type = "Task" [ 1093.752140] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.762050] env[63531]: DEBUG oslo_vmware.api [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118615, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.833028] env[63531]: DEBUG oslo_vmware.api [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118614, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.686167} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.833028] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] 148c6ecd-354e-4076-b52b-c13a2119da7e/148c6ecd-354e-4076-b52b-c13a2119da7e.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1093.833028] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1093.833028] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6fb8a93e-c669-4e05-8335-124d3eaa07f1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.846152] env[63531]: DEBUG oslo_vmware.api [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 1093.846152] env[63531]: value = "task-1118616" [ 1093.846152] env[63531]: _type = "Task" [ 1093.846152] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.855235] env[63531]: DEBUG oslo_vmware.api [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118616, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.860439] env[63531]: DEBUG nova.scheduler.client.report [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1093.925076] env[63531]: DEBUG nova.compute.manager [req-4a187a43-668c-4bd8-abbb-b606e28efb80 req-b220c255-96ef-4440-87aa-cd9316096f5f service nova] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Received event network-vif-plugged-e9692fb2-f3a9-455c-8317-b533a6af48f7 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1093.925323] env[63531]: DEBUG oslo_concurrency.lockutils [req-4a187a43-668c-4bd8-abbb-b606e28efb80 req-b220c255-96ef-4440-87aa-cd9316096f5f service nova] Acquiring lock "ccc8f366-5849-45a6-9745-ff9b2be9c3bb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1093.925533] env[63531]: DEBUG oslo_concurrency.lockutils [req-4a187a43-668c-4bd8-abbb-b606e28efb80 req-b220c255-96ef-4440-87aa-cd9316096f5f service nova] Lock "ccc8f366-5849-45a6-9745-ff9b2be9c3bb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1093.925804] env[63531]: DEBUG oslo_concurrency.lockutils [req-4a187a43-668c-4bd8-abbb-b606e28efb80 req-b220c255-96ef-4440-87aa-cd9316096f5f service nova] Lock "ccc8f366-5849-45a6-9745-ff9b2be9c3bb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1093.925872] env[63531]: DEBUG nova.compute.manager [req-4a187a43-668c-4bd8-abbb-b606e28efb80 req-b220c255-96ef-4440-87aa-cd9316096f5f service nova] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] No waiting events found dispatching network-vif-plugged-e9692fb2-f3a9-455c-8317-b533a6af48f7 {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1093.927448] env[63531]: WARNING nova.compute.manager [req-4a187a43-668c-4bd8-abbb-b606e28efb80 req-b220c255-96ef-4440-87aa-cd9316096f5f service nova] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Received unexpected event network-vif-plugged-e9692fb2-f3a9-455c-8317-b533a6af48f7 for instance with vm_state building and task_state spawning. [ 1094.002113] env[63531]: DEBUG nova.network.neutron [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Successfully updated port: e9692fb2-f3a9-455c-8317-b533a6af48f7 {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1094.264065] env[63531]: DEBUG oslo_vmware.api [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118615, 'name': PowerOffVM_Task, 'duration_secs': 0.245747} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.264961] env[63531]: DEBUG nova.network.neutron [req-0b98624c-4cdc-468e-a53d-8dcb137dc7be req-d940b3aa-34de-4f45-a4f6-8fc53a306f8a service nova] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Updated VIF entry in instance network info cache for port 76060df9-1ff6-4c64-86ef-a8e0d4081cfd. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1094.265310] env[63531]: DEBUG nova.network.neutron [req-0b98624c-4cdc-468e-a53d-8dcb137dc7be req-d940b3aa-34de-4f45-a4f6-8fc53a306f8a service nova] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Updating instance_info_cache with network_info: [{"id": "76060df9-1ff6-4c64-86ef-a8e0d4081cfd", "address": "fa:16:3e:52:f5:96", "network": {"id": "eac8786d-bc67-4916-8483-b53d9a25efb8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1604989690-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "abdae960d43b47bc95ab41f265d2c85b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db068f71-08cc-42d4-8ab6-17134c1585e5", "external-id": "nsx-vlan-transportzone-721", "segmentation_id": 721, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap76060df9-1f", "ovs_interfaceid": "76060df9-1ff6-4c64-86ef-a8e0d4081cfd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1094.266989] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1094.267191] env[63531]: DEBUG nova.virt.hardware [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1094.267414] env[63531]: DEBUG nova.virt.hardware [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1094.267576] env[63531]: DEBUG nova.virt.hardware [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1094.267788] env[63531]: DEBUG nova.virt.hardware [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1094.267956] env[63531]: DEBUG nova.virt.hardware [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1094.268127] env[63531]: DEBUG nova.virt.hardware [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1094.268339] env[63531]: DEBUG nova.virt.hardware [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1094.268503] env[63531]: DEBUG nova.virt.hardware [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1094.268675] env[63531]: DEBUG nova.virt.hardware [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1094.268840] env[63531]: DEBUG nova.virt.hardware [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1094.269034] env[63531]: DEBUG nova.virt.hardware [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1094.274187] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0d68a1bb-2437-4046-8d54-fb6f214d262b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.291548] env[63531]: DEBUG oslo_vmware.api [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 1094.291548] env[63531]: value = "task-1118617" [ 1094.291548] env[63531]: _type = "Task" [ 1094.291548] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.299599] env[63531]: DEBUG oslo_vmware.api [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118617, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.354819] env[63531]: DEBUG oslo_vmware.api [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118616, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076083} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.355107] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1094.355941] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8edb5973-1729-483b-b376-3a92c3cc4b0e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.370049] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63531) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1094.370647] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.832s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1094.378785] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] 148c6ecd-354e-4076-b52b-c13a2119da7e/148c6ecd-354e-4076-b52b-c13a2119da7e.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1094.379126] env[63531]: DEBUG oslo_concurrency.lockutils [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 3.731s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1094.380584] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f9d1df9a-bc75-4d30-ac6d-8fd461ed7538 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.403802] env[63531]: DEBUG oslo_vmware.api [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 1094.403802] env[63531]: value = "task-1118618" [ 1094.403802] env[63531]: _type = "Task" [ 1094.403802] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.418140] env[63531]: DEBUG oslo_vmware.api [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118618, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.507760] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "refresh_cache-ccc8f366-5849-45a6-9745-ff9b2be9c3bb" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1094.507930] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquired lock "refresh_cache-ccc8f366-5849-45a6-9745-ff9b2be9c3bb" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1094.508084] env[63531]: DEBUG nova.network.neutron [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1094.775557] env[63531]: DEBUG oslo_concurrency.lockutils [req-0b98624c-4cdc-468e-a53d-8dcb137dc7be req-d940b3aa-34de-4f45-a4f6-8fc53a306f8a service nova] Releasing lock "refresh_cache-8dec0cac-defc-42f2-ab56-6b3ae60ad858" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1094.801640] env[63531]: DEBUG oslo_vmware.api [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118617, 'name': ReconfigVM_Task, 'duration_secs': 0.159155} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.802487] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-446a0dd7-ddd2-44f5-85f5-e9bbc5f98c11 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.826381] env[63531]: DEBUG nova.virt.hardware [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1094.826671] env[63531]: DEBUG nova.virt.hardware [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1094.826809] env[63531]: DEBUG nova.virt.hardware [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1094.826999] env[63531]: DEBUG nova.virt.hardware [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1094.827172] env[63531]: DEBUG nova.virt.hardware [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1094.827324] env[63531]: DEBUG nova.virt.hardware [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1094.827537] env[63531]: DEBUG nova.virt.hardware [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1094.827705] env[63531]: DEBUG nova.virt.hardware [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1094.827877] env[63531]: DEBUG nova.virt.hardware [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1094.828062] env[63531]: DEBUG nova.virt.hardware [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1094.828245] env[63531]: DEBUG nova.virt.hardware [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1094.829149] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b3a017cc-b623-4296-a6c4-36a3399393bb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.836688] env[63531]: DEBUG oslo_vmware.api [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 1094.836688] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5283931c-8e02-016f-ee09-389befbae774" [ 1094.836688] env[63531]: _type = "Task" [ 1094.836688] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.845578] env[63531]: DEBUG oslo_vmware.api [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5283931c-8e02-016f-ee09-389befbae774, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.896491] env[63531]: DEBUG nova.objects.instance [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lazy-loading 'migration_context' on Instance uuid 96c2910b-1e25-4053-9ad9-c155402ef0ac {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1094.914644] env[63531]: DEBUG oslo_vmware.api [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118618, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.044963] env[63531]: DEBUG nova.network.neutron [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1095.058046] env[63531]: DEBUG nova.compute.manager [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Stashing vm_state: active {{(pid=63531) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1095.198313] env[63531]: DEBUG nova.network.neutron [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Updating instance_info_cache with network_info: [{"id": "e9692fb2-f3a9-455c-8317-b533a6af48f7", "address": "fa:16:3e:79:4a:3b", "network": {"id": "caf86428-59a5-4e50-9135-2ab8473af64d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1673994696-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ba8b284854242c392aec5326e996239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape9692fb2-f3", "ovs_interfaceid": "e9692fb2-f3a9-455c-8317-b533a6af48f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1095.347172] env[63531]: DEBUG oslo_vmware.api [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5283931c-8e02-016f-ee09-389befbae774, 'name': SearchDatastore_Task, 'duration_secs': 0.021672} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.352570] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Reconfiguring VM instance instance-0000005c to detach disk 2000 {{(pid=63531) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1095.352859] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-66e68901-938b-430b-b6d2-94d96792aa09 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.371740] env[63531]: DEBUG oslo_vmware.api [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 1095.371740] env[63531]: value = "task-1118620" [ 1095.371740] env[63531]: _type = "Task" [ 1095.371740] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.379805] env[63531]: DEBUG oslo_vmware.api [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118620, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.413947] env[63531]: DEBUG oslo_vmware.api [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118618, 'name': ReconfigVM_Task, 'duration_secs': 0.830177} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.414306] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] Reconfigured VM instance instance-0000006c to attach disk [datastore2] 148c6ecd-354e-4076-b52b-c13a2119da7e/148c6ecd-354e-4076-b52b-c13a2119da7e.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1095.415150] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b5bc1fdd-3052-4221-b99d-2bb20066cddd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.423557] env[63531]: DEBUG oslo_vmware.api [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 1095.423557] env[63531]: value = "task-1118621" [ 1095.423557] env[63531]: _type = "Task" [ 1095.423557] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.433446] env[63531]: DEBUG oslo_vmware.api [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118621, 'name': Rename_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.528386] env[63531]: DEBUG nova.compute.manager [req-b52dbe13-5dd6-4c54-bec7-9aee048771f6 req-498ae106-7ae1-4275-b4f0-7459e53ccef2 service nova] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Received event network-changed-99a45e7c-1ddb-4c35-ac69-b2c33f976019 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1095.528566] env[63531]: DEBUG nova.compute.manager [req-b52dbe13-5dd6-4c54-bec7-9aee048771f6 req-498ae106-7ae1-4275-b4f0-7459e53ccef2 service nova] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Refreshing instance network info cache due to event network-changed-99a45e7c-1ddb-4c35-ac69-b2c33f976019. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1095.528793] env[63531]: DEBUG oslo_concurrency.lockutils [req-b52dbe13-5dd6-4c54-bec7-9aee048771f6 req-498ae106-7ae1-4275-b4f0-7459e53ccef2 service nova] Acquiring lock "refresh_cache-49967454-a4f9-47f6-a2a9-0cd0c29e18d8" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1095.528941] env[63531]: DEBUG oslo_concurrency.lockutils [req-b52dbe13-5dd6-4c54-bec7-9aee048771f6 req-498ae106-7ae1-4275-b4f0-7459e53ccef2 service nova] Acquired lock "refresh_cache-49967454-a4f9-47f6-a2a9-0cd0c29e18d8" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1095.529120] env[63531]: DEBUG nova.network.neutron [req-b52dbe13-5dd6-4c54-bec7-9aee048771f6 req-498ae106-7ae1-4275-b4f0-7459e53ccef2 service nova] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Refreshing network info cache for port 99a45e7c-1ddb-4c35-ac69-b2c33f976019 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1095.569273] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-229d5d75-155c-4b9d-ba40-21a48eb7fec2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.577386] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1095.578512] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ce98e4b-38f8-41f3-83da-8a822f96bbbb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.609617] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8450e87-119e-443e-8ea5-8b7d1bf0ff2d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.617677] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21357892-cefe-44f2-ada8-43e16a593b24 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.630726] env[63531]: DEBUG nova.compute.provider_tree [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1095.700857] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Releasing lock "refresh_cache-ccc8f366-5849-45a6-9745-ff9b2be9c3bb" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1095.701201] env[63531]: DEBUG nova.compute.manager [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Instance network_info: |[{"id": "e9692fb2-f3a9-455c-8317-b533a6af48f7", "address": "fa:16:3e:79:4a:3b", "network": {"id": "caf86428-59a5-4e50-9135-2ab8473af64d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1673994696-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ba8b284854242c392aec5326e996239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape9692fb2-f3", "ovs_interfaceid": "e9692fb2-f3a9-455c-8317-b533a6af48f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1095.701620] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:79:4a:3b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '91b0f7e5-0d1a-46e2-bf73-09656211dea2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e9692fb2-f3a9-455c-8317-b533a6af48f7', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1095.709068] env[63531]: DEBUG oslo.service.loopingcall [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1095.709656] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1095.709899] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-32258e5c-da85-4f18-8407-3daf821197cb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.730821] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1095.730821] env[63531]: value = "task-1118622" [ 1095.730821] env[63531]: _type = "Task" [ 1095.730821] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.743640] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118622, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.884171] env[63531]: DEBUG oslo_vmware.api [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118620, 'name': ReconfigVM_Task, 'duration_secs': 0.245891} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.884503] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Reconfigured VM instance instance-0000005c to detach disk 2000 {{(pid=63531) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1095.885246] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d3d3be8-bae4-491d-a1d0-b80f46b4d10c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.910309] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Reconfiguring VM instance instance-0000005c to attach disk [datastore1] 18057005-1ff1-4e3d-bccc-8c89755d6db7/18057005-1ff1-4e3d-bccc-8c89755d6db7.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1095.910570] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7113c7af-30b5-4657-86ef-c58494a8c880 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.933972] env[63531]: DEBUG oslo_vmware.api [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118621, 'name': Rename_Task, 'duration_secs': 0.144667} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.934290] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1095.934564] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-99d679c5-0534-4baa-8fb7-64199dabc4bb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.937338] env[63531]: DEBUG oslo_vmware.api [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 1095.937338] env[63531]: value = "task-1118623" [ 1095.937338] env[63531]: _type = "Task" [ 1095.937338] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.943509] env[63531]: DEBUG oslo_vmware.api [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 1095.943509] env[63531]: value = "task-1118624" [ 1095.943509] env[63531]: _type = "Task" [ 1095.943509] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.946934] env[63531]: DEBUG oslo_vmware.api [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118623, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.952835] env[63531]: DEBUG nova.compute.manager [req-785ec5c7-94d8-46a7-a704-a21819fd6a2f req-3cd5c48b-1cfb-4234-b891-c363018194b4 service nova] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Received event network-changed-e9692fb2-f3a9-455c-8317-b533a6af48f7 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1095.952920] env[63531]: DEBUG nova.compute.manager [req-785ec5c7-94d8-46a7-a704-a21819fd6a2f req-3cd5c48b-1cfb-4234-b891-c363018194b4 service nova] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Refreshing instance network info cache due to event network-changed-e9692fb2-f3a9-455c-8317-b533a6af48f7. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1095.953154] env[63531]: DEBUG oslo_concurrency.lockutils [req-785ec5c7-94d8-46a7-a704-a21819fd6a2f req-3cd5c48b-1cfb-4234-b891-c363018194b4 service nova] Acquiring lock "refresh_cache-ccc8f366-5849-45a6-9745-ff9b2be9c3bb" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1095.953305] env[63531]: DEBUG oslo_concurrency.lockutils [req-785ec5c7-94d8-46a7-a704-a21819fd6a2f req-3cd5c48b-1cfb-4234-b891-c363018194b4 service nova] Acquired lock "refresh_cache-ccc8f366-5849-45a6-9745-ff9b2be9c3bb" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1095.953470] env[63531]: DEBUG nova.network.neutron [req-785ec5c7-94d8-46a7-a704-a21819fd6a2f req-3cd5c48b-1cfb-4234-b891-c363018194b4 service nova] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Refreshing network info cache for port e9692fb2-f3a9-455c-8317-b533a6af48f7 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1095.958825] env[63531]: DEBUG oslo_vmware.api [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118624, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.134790] env[63531]: DEBUG nova.scheduler.client.report [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1096.243476] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118622, 'name': CreateVM_Task, 'duration_secs': 0.333428} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.243669] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1096.244385] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1096.244556] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1096.244879] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1096.245150] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d130862c-69bb-4ae0-9bf9-37521c7ca871 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.250213] env[63531]: DEBUG oslo_vmware.api [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for the task: (returnval){ [ 1096.250213] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52e97c3d-47c5-47f3-c867-14899227a41f" [ 1096.250213] env[63531]: _type = "Task" [ 1096.250213] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.258190] env[63531]: DEBUG oslo_vmware.api [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52e97c3d-47c5-47f3-c867-14899227a41f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.258978] env[63531]: DEBUG nova.network.neutron [req-b52dbe13-5dd6-4c54-bec7-9aee048771f6 req-498ae106-7ae1-4275-b4f0-7459e53ccef2 service nova] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Updated VIF entry in instance network info cache for port 99a45e7c-1ddb-4c35-ac69-b2c33f976019. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1096.259349] env[63531]: DEBUG nova.network.neutron [req-b52dbe13-5dd6-4c54-bec7-9aee048771f6 req-498ae106-7ae1-4275-b4f0-7459e53ccef2 service nova] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Updating instance_info_cache with network_info: [{"id": "99a45e7c-1ddb-4c35-ac69-b2c33f976019", "address": "fa:16:3e:c8:19:0d", "network": {"id": "eac8786d-bc67-4916-8483-b53d9a25efb8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1604989690-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.241", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "abdae960d43b47bc95ab41f265d2c85b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db068f71-08cc-42d4-8ab6-17134c1585e5", "external-id": "nsx-vlan-transportzone-721", "segmentation_id": 721, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap99a45e7c-1d", "ovs_interfaceid": "99a45e7c-1ddb-4c35-ac69-b2c33f976019", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1096.448925] env[63531]: DEBUG oslo_vmware.api [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118623, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.456969] env[63531]: DEBUG oslo_vmware.api [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118624, 'name': PowerOnVM_Task, 'duration_secs': 0.48511} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.457254] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1096.457464] env[63531]: INFO nova.compute.manager [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] Took 9.53 seconds to spawn the instance on the hypervisor. [ 1096.457660] env[63531]: DEBUG nova.compute.manager [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1096.458384] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef61bf84-2a66-40de-af61-2ba15fc9a77f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.649188] env[63531]: DEBUG nova.network.neutron [req-785ec5c7-94d8-46a7-a704-a21819fd6a2f req-3cd5c48b-1cfb-4234-b891-c363018194b4 service nova] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Updated VIF entry in instance network info cache for port e9692fb2-f3a9-455c-8317-b533a6af48f7. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1096.649340] env[63531]: DEBUG nova.network.neutron [req-785ec5c7-94d8-46a7-a704-a21819fd6a2f req-3cd5c48b-1cfb-4234-b891-c363018194b4 service nova] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Updating instance_info_cache with network_info: [{"id": "e9692fb2-f3a9-455c-8317-b533a6af48f7", "address": "fa:16:3e:79:4a:3b", "network": {"id": "caf86428-59a5-4e50-9135-2ab8473af64d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1673994696-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ba8b284854242c392aec5326e996239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape9692fb2-f3", "ovs_interfaceid": "e9692fb2-f3a9-455c-8317-b533a6af48f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1096.762632] env[63531]: DEBUG oslo_concurrency.lockutils [req-b52dbe13-5dd6-4c54-bec7-9aee048771f6 req-498ae106-7ae1-4275-b4f0-7459e53ccef2 service nova] Releasing lock "refresh_cache-49967454-a4f9-47f6-a2a9-0cd0c29e18d8" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1096.762986] env[63531]: DEBUG oslo_vmware.api [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52e97c3d-47c5-47f3-c867-14899227a41f, 'name': SearchDatastore_Task, 'duration_secs': 0.014381} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.763283] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1096.763511] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1096.763748] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1096.763898] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1096.764095] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1096.764399] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0ec19baa-a289-4069-bff0-aa0a41c11b38 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.774644] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1096.774824] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1096.775524] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bad871d2-1ab1-476d-9731-0c48063b1f35 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.780606] env[63531]: DEBUG oslo_vmware.api [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for the task: (returnval){ [ 1096.780606] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52fa6fda-128e-450c-dc90-fc247e0173c4" [ 1096.780606] env[63531]: _type = "Task" [ 1096.780606] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.788509] env[63531]: DEBUG oslo_vmware.api [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52fa6fda-128e-450c-dc90-fc247e0173c4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.949366] env[63531]: DEBUG oslo_vmware.api [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118623, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.976824] env[63531]: INFO nova.compute.manager [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] Took 20.02 seconds to build instance. [ 1097.148109] env[63531]: DEBUG oslo_concurrency.lockutils [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.769s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1097.157794] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.016s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1097.159892] env[63531]: INFO nova.compute.claims [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] [instance: fbf63fa0-c8ec-41c7-878e-1446537b7b49] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1097.163495] env[63531]: DEBUG oslo_concurrency.lockutils [req-785ec5c7-94d8-46a7-a704-a21819fd6a2f req-3cd5c48b-1cfb-4234-b891-c363018194b4 service nova] Releasing lock "refresh_cache-ccc8f366-5849-45a6-9745-ff9b2be9c3bb" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1097.291617] env[63531]: DEBUG oslo_vmware.api [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52fa6fda-128e-450c-dc90-fc247e0173c4, 'name': SearchDatastore_Task, 'duration_secs': 0.033741} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.292516] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6c1249e8-1c57-49d1-bd04-0f6d9526cd8e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.298427] env[63531]: DEBUG oslo_vmware.api [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for the task: (returnval){ [ 1097.298427] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52dae510-3d0b-9f1e-d763-4f7260618a3b" [ 1097.298427] env[63531]: _type = "Task" [ 1097.298427] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.307386] env[63531]: DEBUG oslo_vmware.api [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52dae510-3d0b-9f1e-d763-4f7260618a3b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.450391] env[63531]: DEBUG oslo_vmware.api [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118623, 'name': ReconfigVM_Task, 'duration_secs': 1.308382} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.450684] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Reconfigured VM instance instance-0000005c to attach disk [datastore1] 18057005-1ff1-4e3d-bccc-8c89755d6db7/18057005-1ff1-4e3d-bccc-8c89755d6db7.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1097.451623] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db7ed3c4-ffea-42fa-a57b-d8302c5b6a86 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.475905] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10648de1-ec2d-42db-b8ab-1de7e48c4713 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.478696] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bf94bc8b-906b-41f4-b54b-2de29575a988 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "148c6ecd-354e-4076-b52b-c13a2119da7e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.532s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1097.498740] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3066375-c9f2-4ba7-8ff1-f73ff897f0a7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.521405] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f22f974-d1ff-4766-a6c2-8361ef09e4ae {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.528856] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1097.529134] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-47d6e993-c923-425b-bdd0-015ac19a6d79 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.536442] env[63531]: DEBUG oslo_vmware.api [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 1097.536442] env[63531]: value = "task-1118625" [ 1097.536442] env[63531]: _type = "Task" [ 1097.536442] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.544425] env[63531]: DEBUG oslo_vmware.api [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118625, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.810215] env[63531]: DEBUG oslo_vmware.api [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52dae510-3d0b-9f1e-d763-4f7260618a3b, 'name': SearchDatastore_Task, 'duration_secs': 0.010338} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.810215] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1097.810215] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] ccc8f366-5849-45a6-9745-ff9b2be9c3bb/ccc8f366-5849-45a6-9745-ff9b2be9c3bb.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1097.810215] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c8f2f6ad-d019-4d5f-af14-c5a93f8c6225 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.816902] env[63531]: DEBUG oslo_vmware.api [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for the task: (returnval){ [ 1097.816902] env[63531]: value = "task-1118626" [ 1097.816902] env[63531]: _type = "Task" [ 1097.816902] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.824715] env[63531]: DEBUG oslo_vmware.api [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118626, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.937453] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e873d745-b972-4101-969a-4b6054ead779 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.944679] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-468ef3ea-e470-4387-9625-68a8f1c43f93 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] Suspending the VM {{(pid=63531) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 1097.944972] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-b22eb7a9-c6a2-4097-b653-ac8dce7a827c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.952019] env[63531]: DEBUG oslo_vmware.api [None req-468ef3ea-e470-4387-9625-68a8f1c43f93 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 1097.952019] env[63531]: value = "task-1118627" [ 1097.952019] env[63531]: _type = "Task" [ 1097.952019] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.961057] env[63531]: DEBUG oslo_vmware.api [None req-468ef3ea-e470-4387-9625-68a8f1c43f93 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118627, 'name': SuspendVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.046949] env[63531]: DEBUG oslo_vmware.api [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118625, 'name': PowerOnVM_Task, 'duration_secs': 0.408232} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.047329] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1098.329392] env[63531]: DEBUG oslo_vmware.api [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118626, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.393462] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc2e97dd-e44d-441f-afaf-bfe61516a0ca {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.404237] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22b91ac4-d50a-43ee-b3f0-c26f2344bd8f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.438070] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbc414ec-c56c-4a40-8341-5d7436746ecb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.446747] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29bce731-8926-4baa-be17-0cf709dcb50b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.467695] env[63531]: DEBUG nova.compute.provider_tree [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1098.477810] env[63531]: DEBUG oslo_vmware.api [None req-468ef3ea-e470-4387-9625-68a8f1c43f93 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118627, 'name': SuspendVM_Task} progress is 91%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.694728] env[63531]: INFO nova.compute.manager [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Swapping old allocation on dict_keys(['9996a314-a62e-4282-8252-f387e7e3fabb']) held by migration 063e7061-74ea-430a-93f4-e2b436fafba7 for instance [ 1098.720846] env[63531]: DEBUG nova.scheduler.client.report [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Overwriting current allocation {'allocations': {'9996a314-a62e-4282-8252-f387e7e3fabb': {'resources': {'VCPU': 1, 'MEMORY_MB': 256, 'DISK_GB': 1}, 'generation': 149}}, 'project_id': '308eabafea9c4188a58a0f1c22074d2f', 'user_id': '597d424341124b7db23dc7a104107148', 'consumer_generation': 1} on consumer 96c2910b-1e25-4053-9ad9-c155402ef0ac {{(pid=63531) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1098.830717] env[63531]: DEBUG oslo_vmware.api [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118626, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.762084} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.831875] env[63531]: DEBUG oslo_concurrency.lockutils [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "refresh_cache-96c2910b-1e25-4053-9ad9-c155402ef0ac" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1098.832269] env[63531]: DEBUG oslo_concurrency.lockutils [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquired lock "refresh_cache-96c2910b-1e25-4053-9ad9-c155402ef0ac" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1098.832479] env[63531]: DEBUG nova.network.neutron [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1098.833834] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] ccc8f366-5849-45a6-9745-ff9b2be9c3bb/ccc8f366-5849-45a6-9745-ff9b2be9c3bb.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1098.834632] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1098.835344] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-db447300-7299-4e58-87e7-09c33dea05cd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.843865] env[63531]: DEBUG oslo_vmware.api [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for the task: (returnval){ [ 1098.843865] env[63531]: value = "task-1118628" [ 1098.843865] env[63531]: _type = "Task" [ 1098.843865] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.853899] env[63531]: DEBUG oslo_vmware.api [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118628, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.965384] env[63531]: DEBUG oslo_vmware.api [None req-468ef3ea-e470-4387-9625-68a8f1c43f93 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118627, 'name': SuspendVM_Task, 'duration_secs': 0.918704} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.965768] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-468ef3ea-e470-4387-9625-68a8f1c43f93 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] Suspended the VM {{(pid=63531) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 1098.965843] env[63531]: DEBUG nova.compute.manager [None req-468ef3ea-e470-4387-9625-68a8f1c43f93 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1098.966713] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56356966-7402-4bd8-a2db-2eea08e5f78f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.974354] env[63531]: DEBUG nova.scheduler.client.report [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1099.098179] env[63531]: INFO nova.compute.manager [None req-40fddbb5-272a-4928-a4e5-837f428cdd68 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Updating instance to original state: 'active' [ 1099.355536] env[63531]: DEBUG oslo_vmware.api [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118628, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073301} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.355844] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1099.356642] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e70e12f7-f83b-413e-a937-38deb12dcb94 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.380274] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Reconfiguring VM instance instance-0000006d to attach disk [datastore2] ccc8f366-5849-45a6-9745-ff9b2be9c3bb/ccc8f366-5849-45a6-9745-ff9b2be9c3bb.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1099.383446] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-51e348f8-70cd-41d8-a89a-d909eb4c60e0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.406356] env[63531]: DEBUG oslo_vmware.api [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for the task: (returnval){ [ 1099.406356] env[63531]: value = "task-1118629" [ 1099.406356] env[63531]: _type = "Task" [ 1099.406356] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.416528] env[63531]: DEBUG oslo_vmware.api [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118629, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.483433] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.326s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1099.483911] env[63531]: DEBUG nova.compute.manager [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] [instance: fbf63fa0-c8ec-41c7-878e-1446537b7b49] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1099.491942] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a13a31cb-b684-4c2f-ae54-622a7c633f10 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.569s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1099.491942] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a13a31cb-b684-4c2f-ae54-622a7c633f10 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1099.491942] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 3.914s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1099.516295] env[63531]: INFO nova.scheduler.client.report [None req-a13a31cb-b684-4c2f-ae54-622a7c633f10 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Deleted allocations for instance b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16 [ 1099.625197] env[63531]: DEBUG nova.network.neutron [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Updating instance_info_cache with network_info: [{"id": "b731766f-738c-4d77-a00e-98ed6c683e55", "address": "fa:16:3e:d5:26:02", "network": {"id": "0667c176-cfdf-4e43-b152-e00bdc224448", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-243715985-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.217", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "308eabafea9c4188a58a0f1c22074d2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8f580e6-1d86-41ee-9ebe-c531cb9299c6", "external-id": "nsx-vlan-transportzone-150", "segmentation_id": 150, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb731766f-73", "ovs_interfaceid": "b731766f-738c-4d77-a00e-98ed6c683e55", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1099.918063] env[63531]: DEBUG oslo_vmware.api [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118629, 'name': ReconfigVM_Task, 'duration_secs': 0.296759} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.918828] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Reconfigured VM instance instance-0000006d to attach disk [datastore2] ccc8f366-5849-45a6-9745-ff9b2be9c3bb/ccc8f366-5849-45a6-9745-ff9b2be9c3bb.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1099.919749] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c5fe77e6-4771-436e-bbd9-f7478e1d738c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.938211] env[63531]: DEBUG oslo_vmware.api [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for the task: (returnval){ [ 1099.938211] env[63531]: value = "task-1118630" [ 1099.938211] env[63531]: _type = "Task" [ 1099.938211] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.954391] env[63531]: DEBUG oslo_vmware.api [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118630, 'name': Rename_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.995446] env[63531]: DEBUG nova.compute.utils [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1099.998743] env[63531]: INFO nova.compute.claims [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1100.001981] env[63531]: DEBUG nova.compute.manager [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] [instance: fbf63fa0-c8ec-41c7-878e-1446537b7b49] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1100.002411] env[63531]: DEBUG nova.network.neutron [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] [instance: fbf63fa0-c8ec-41c7-878e-1446537b7b49] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1100.026216] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a13a31cb-b684-4c2f-ae54-622a7c633f10 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.123s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1100.067726] env[63531]: DEBUG nova.policy [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f77f45233e4649c8a9d05397f48c1bff', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '126f5465d9be40bf8b6d1dff02f10c7a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 1100.128324] env[63531]: DEBUG oslo_concurrency.lockutils [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Releasing lock "refresh_cache-96c2910b-1e25-4053-9ad9-c155402ef0ac" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1100.128797] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1100.129155] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8d08d531-52a4-44e1-89c6-4bf9e532fb6f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.138499] env[63531]: DEBUG oslo_vmware.api [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 1100.138499] env[63531]: value = "task-1118631" [ 1100.138499] env[63531]: _type = "Task" [ 1100.138499] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.149224] env[63531]: DEBUG oslo_vmware.api [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118631, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.343467] env[63531]: DEBUG nova.network.neutron [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] [instance: fbf63fa0-c8ec-41c7-878e-1446537b7b49] Successfully created port: 4fa0186c-186b-4f32-8db3-dc07235c8803 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1100.448437] env[63531]: DEBUG oslo_vmware.api [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118630, 'name': Rename_Task, 'duration_secs': 0.157534} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.448751] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1100.449061] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d3559983-fdc1-4613-9ad7-b42cd642d487 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.456759] env[63531]: DEBUG oslo_vmware.api [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for the task: (returnval){ [ 1100.456759] env[63531]: value = "task-1118632" [ 1100.456759] env[63531]: _type = "Task" [ 1100.456759] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.468135] env[63531]: DEBUG oslo_vmware.api [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118632, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.475829] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c4d0011c-eb80-4408-9de7-9a716bf764ba tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "148c6ecd-354e-4076-b52b-c13a2119da7e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1100.476196] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c4d0011c-eb80-4408-9de7-9a716bf764ba tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "148c6ecd-354e-4076-b52b-c13a2119da7e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1100.476419] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c4d0011c-eb80-4408-9de7-9a716bf764ba tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "148c6ecd-354e-4076-b52b-c13a2119da7e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1100.476612] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c4d0011c-eb80-4408-9de7-9a716bf764ba tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "148c6ecd-354e-4076-b52b-c13a2119da7e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1100.476785] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c4d0011c-eb80-4408-9de7-9a716bf764ba tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "148c6ecd-354e-4076-b52b-c13a2119da7e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1100.479322] env[63531]: INFO nova.compute.manager [None req-c4d0011c-eb80-4408-9de7-9a716bf764ba tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] Terminating instance [ 1100.481263] env[63531]: DEBUG nova.compute.manager [None req-c4d0011c-eb80-4408-9de7-9a716bf764ba tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1100.481458] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c4d0011c-eb80-4408-9de7-9a716bf764ba tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1100.482288] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aa3a816-6d2b-4d35-b897-7ef665022ae6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.494569] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c4d0011c-eb80-4408-9de7-9a716bf764ba tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1100.494874] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1acfdc84-6f6c-4e2f-95f8-055993e06601 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.510024] env[63531]: INFO nova.compute.resource_tracker [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Updating resource usage from migration 72e41fa6-b1eb-48d8-9701-d5fc7cd368c7 [ 1100.513090] env[63531]: DEBUG nova.compute.manager [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] [instance: fbf63fa0-c8ec-41c7-878e-1446537b7b49] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1100.585993] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c4d0011c-eb80-4408-9de7-9a716bf764ba tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1100.586668] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c4d0011c-eb80-4408-9de7-9a716bf764ba tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] Deleting contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1100.586781] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4d0011c-eb80-4408-9de7-9a716bf764ba tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Deleting the datastore file [datastore2] 148c6ecd-354e-4076-b52b-c13a2119da7e {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1100.587150] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-668efd9e-2253-4880-aaf3-c1a70d21eddb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.596030] env[63531]: DEBUG oslo_vmware.api [None req-c4d0011c-eb80-4408-9de7-9a716bf764ba tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 1100.596030] env[63531]: value = "task-1118634" [ 1100.596030] env[63531]: _type = "Task" [ 1100.596030] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.605771] env[63531]: DEBUG oslo_vmware.api [None req-c4d0011c-eb80-4408-9de7-9a716bf764ba tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118634, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.654207] env[63531]: DEBUG oslo_vmware.api [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118631, 'name': PowerOffVM_Task, 'duration_secs': 0.214655} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.654636] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1100.655464] env[63531]: DEBUG nova.virt.hardware [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1100.655711] env[63531]: DEBUG nova.virt.hardware [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1100.655881] env[63531]: DEBUG nova.virt.hardware [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1100.656083] env[63531]: DEBUG nova.virt.hardware [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1100.656255] env[63531]: DEBUG nova.virt.hardware [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1100.656413] env[63531]: DEBUG nova.virt.hardware [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1100.656623] env[63531]: DEBUG nova.virt.hardware [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1100.656799] env[63531]: DEBUG nova.virt.hardware [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1100.657054] env[63531]: DEBUG nova.virt.hardware [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1100.657277] env[63531]: DEBUG nova.virt.hardware [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1100.657605] env[63531]: DEBUG nova.virt.hardware [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1100.666321] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-442f4104-5f53-4e23-bc11-d9c94799945e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.691590] env[63531]: DEBUG oslo_vmware.api [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 1100.691590] env[63531]: value = "task-1118635" [ 1100.691590] env[63531]: _type = "Task" [ 1100.691590] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.704921] env[63531]: DEBUG oslo_vmware.api [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118635, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.718295] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b57161f-c814-4771-8536-b07acbcb486e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.726540] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-472e0dbc-5b7c-49eb-87fa-1775643b5d15 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.768549] env[63531]: DEBUG oslo_concurrency.lockutils [None req-de477bf0-a47e-4909-bd9d-2464242d52bb tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquiring lock "18057005-1ff1-4e3d-bccc-8c89755d6db7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1100.768848] env[63531]: DEBUG oslo_concurrency.lockutils [None req-de477bf0-a47e-4909-bd9d-2464242d52bb tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "18057005-1ff1-4e3d-bccc-8c89755d6db7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1100.769101] env[63531]: DEBUG oslo_concurrency.lockutils [None req-de477bf0-a47e-4909-bd9d-2464242d52bb tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquiring lock "18057005-1ff1-4e3d-bccc-8c89755d6db7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1100.769310] env[63531]: DEBUG oslo_concurrency.lockutils [None req-de477bf0-a47e-4909-bd9d-2464242d52bb tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "18057005-1ff1-4e3d-bccc-8c89755d6db7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1100.769506] env[63531]: DEBUG oslo_concurrency.lockutils [None req-de477bf0-a47e-4909-bd9d-2464242d52bb tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "18057005-1ff1-4e3d-bccc-8c89755d6db7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1100.772126] env[63531]: INFO nova.compute.manager [None req-de477bf0-a47e-4909-bd9d-2464242d52bb tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Terminating instance [ 1100.774024] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f4a79ba-964b-4531-a3c5-6e8b3c9b3476 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.777228] env[63531]: DEBUG nova.compute.manager [None req-de477bf0-a47e-4909-bd9d-2464242d52bb tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1100.777458] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-de477bf0-a47e-4909-bd9d-2464242d52bb tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1100.777702] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9e7184e3-94d6-4c98-b03c-c455a5b8a13c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.788245] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81bb8ae7-e709-45b7-8b59-c15e9b504931 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.791440] env[63531]: DEBUG oslo_vmware.api [None req-de477bf0-a47e-4909-bd9d-2464242d52bb tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 1100.791440] env[63531]: value = "task-1118636" [ 1100.791440] env[63531]: _type = "Task" [ 1100.791440] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.803025] env[63531]: DEBUG nova.compute.provider_tree [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Updating inventory in ProviderTree for provider 9996a314-a62e-4282-8252-f387e7e3fabb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1100.809412] env[63531]: DEBUG oslo_vmware.api [None req-de477bf0-a47e-4909-bd9d-2464242d52bb tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118636, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.968096] env[63531]: DEBUG oslo_vmware.api [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118632, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.105773] env[63531]: DEBUG oslo_vmware.api [None req-c4d0011c-eb80-4408-9de7-9a716bf764ba tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118634, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.229345} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.106054] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4d0011c-eb80-4408-9de7-9a716bf764ba tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1101.106249] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c4d0011c-eb80-4408-9de7-9a716bf764ba tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] Deleted contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1101.106459] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c4d0011c-eb80-4408-9de7-9a716bf764ba tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1101.106681] env[63531]: INFO nova.compute.manager [None req-c4d0011c-eb80-4408-9de7-9a716bf764ba tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] Took 0.63 seconds to destroy the instance on the hypervisor. [ 1101.106930] env[63531]: DEBUG oslo.service.loopingcall [None req-c4d0011c-eb80-4408-9de7-9a716bf764ba tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1101.107615] env[63531]: DEBUG nova.compute.manager [-] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1101.107726] env[63531]: DEBUG nova.network.neutron [-] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1101.205012] env[63531]: DEBUG oslo_vmware.api [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118635, 'name': ReconfigVM_Task, 'duration_secs': 0.187899} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.205931] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f549f07-7277-45e0-8b55-c95434116b14 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.226399] env[63531]: DEBUG nova.virt.hardware [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1101.226640] env[63531]: DEBUG nova.virt.hardware [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1101.226805] env[63531]: DEBUG nova.virt.hardware [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1101.226998] env[63531]: DEBUG nova.virt.hardware [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1101.227168] env[63531]: DEBUG nova.virt.hardware [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1101.227316] env[63531]: DEBUG nova.virt.hardware [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1101.227524] env[63531]: DEBUG nova.virt.hardware [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1101.227686] env[63531]: DEBUG nova.virt.hardware [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1101.227854] env[63531]: DEBUG nova.virt.hardware [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1101.228030] env[63531]: DEBUG nova.virt.hardware [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1101.228217] env[63531]: DEBUG nova.virt.hardware [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1101.229012] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d676bff4-e228-4bd5-a18c-918d884ec3fb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.235316] env[63531]: DEBUG oslo_vmware.api [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 1101.235316] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]527623a6-bb90-ce4a-4f86-156fc33547f9" [ 1101.235316] env[63531]: _type = "Task" [ 1101.235316] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.244429] env[63531]: DEBUG oslo_vmware.api [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]527623a6-bb90-ce4a-4f86-156fc33547f9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.301391] env[63531]: DEBUG oslo_vmware.api [None req-de477bf0-a47e-4909-bd9d-2464242d52bb tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118636, 'name': PowerOffVM_Task, 'duration_secs': 0.240606} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.301689] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-de477bf0-a47e-4909-bd9d-2464242d52bb tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1101.301915] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-de477bf0-a47e-4909-bd9d-2464242d52bb tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Volume detach. Driver type: vmdk {{(pid=63531) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1101.302181] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-de477bf0-a47e-4909-bd9d-2464242d52bb tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244792', 'volume_id': '5c5f7c58-05a8-42fc-9d14-8298686c8120', 'name': 'volume-5c5f7c58-05a8-42fc-9d14-8298686c8120', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': '18057005-1ff1-4e3d-bccc-8c89755d6db7', 'attached_at': '2024-10-03T08:25:29.000000', 'detached_at': '', 'volume_id': '5c5f7c58-05a8-42fc-9d14-8298686c8120', 'serial': '5c5f7c58-05a8-42fc-9d14-8298686c8120'} {{(pid=63531) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1101.303067] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddb5fe65-9890-48f8-8b34-eac441fbed7b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.329628] env[63531]: ERROR nova.scheduler.client.report [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [req-5035c8ed-90db-43e4-b342-babe0c0cabb6] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 9996a314-a62e-4282-8252-f387e7e3fabb. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-5035c8ed-90db-43e4-b342-babe0c0cabb6"}]} [ 1101.330558] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e231ba0e-97c1-43e3-ae5c-a7b29882ba92 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.340781] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9db0736a-5da1-4107-ab42-ac6dd6977f5d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.362807] env[63531]: DEBUG nova.scheduler.client.report [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Refreshing inventories for resource provider 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1101.365313] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57d1d46c-74d7-4191-b7c2-80c3db438baf {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.382436] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-de477bf0-a47e-4909-bd9d-2464242d52bb tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] The volume has not been displaced from its original location: [datastore2] volume-5c5f7c58-05a8-42fc-9d14-8298686c8120/volume-5c5f7c58-05a8-42fc-9d14-8298686c8120.vmdk. No consolidation needed. {{(pid=63531) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1101.388257] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-de477bf0-a47e-4909-bd9d-2464242d52bb tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Reconfiguring VM instance instance-0000005c to detach disk 2001 {{(pid=63531) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1101.390554] env[63531]: DEBUG nova.scheduler.client.report [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Updating ProviderTree inventory for provider 9996a314-a62e-4282-8252-f387e7e3fabb from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1101.390794] env[63531]: DEBUG nova.compute.provider_tree [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Updating inventory in ProviderTree for provider 9996a314-a62e-4282-8252-f387e7e3fabb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1101.392748] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7f4d438a-e273-405c-969d-1527b82c5a5c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.409531] env[63531]: DEBUG nova.compute.manager [req-94d1f472-8d0b-4cff-8a0a-90ca8df0bb97 req-f03f6d49-0e06-48bf-8014-89f2c82e571e service nova] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] Received event network-vif-deleted-1765d3a3-d6ed-4150-aedf-45d90978133b {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1101.409759] env[63531]: INFO nova.compute.manager [req-94d1f472-8d0b-4cff-8a0a-90ca8df0bb97 req-f03f6d49-0e06-48bf-8014-89f2c82e571e service nova] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] Neutron deleted interface 1765d3a3-d6ed-4150-aedf-45d90978133b; detaching it from the instance and deleting it from the info cache [ 1101.409913] env[63531]: DEBUG nova.network.neutron [req-94d1f472-8d0b-4cff-8a0a-90ca8df0bb97 req-f03f6d49-0e06-48bf-8014-89f2c82e571e service nova] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1101.418450] env[63531]: DEBUG oslo_vmware.api [None req-de477bf0-a47e-4909-bd9d-2464242d52bb tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 1101.418450] env[63531]: value = "task-1118637" [ 1101.418450] env[63531]: _type = "Task" [ 1101.418450] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.419423] env[63531]: DEBUG nova.scheduler.client.report [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Refreshing aggregate associations for resource provider 9996a314-a62e-4282-8252-f387e7e3fabb, aggregates: None {{(pid=63531) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1101.435173] env[63531]: DEBUG oslo_vmware.api [None req-de477bf0-a47e-4909-bd9d-2464242d52bb tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118637, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.441615] env[63531]: DEBUG nova.scheduler.client.report [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Refreshing trait associations for resource provider 9996a314-a62e-4282-8252-f387e7e3fabb, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE {{(pid=63531) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1101.471036] env[63531]: DEBUG oslo_vmware.api [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118632, 'name': PowerOnVM_Task, 'duration_secs': 0.592882} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.473653] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1101.473878] env[63531]: INFO nova.compute.manager [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Took 7.91 seconds to spawn the instance on the hypervisor. [ 1101.474095] env[63531]: DEBUG nova.compute.manager [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1101.475119] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-834bb90e-0123-4c7a-84bf-bf169a8379c0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.524543] env[63531]: DEBUG nova.compute.manager [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] [instance: fbf63fa0-c8ec-41c7-878e-1446537b7b49] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1101.556465] env[63531]: DEBUG nova.virt.hardware [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1101.556750] env[63531]: DEBUG nova.virt.hardware [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1101.556939] env[63531]: DEBUG nova.virt.hardware [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1101.557148] env[63531]: DEBUG nova.virt.hardware [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1101.557306] env[63531]: DEBUG nova.virt.hardware [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1101.557459] env[63531]: DEBUG nova.virt.hardware [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1101.557893] env[63531]: DEBUG nova.virt.hardware [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1101.558088] env[63531]: DEBUG nova.virt.hardware [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1101.558273] env[63531]: DEBUG nova.virt.hardware [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1101.558544] env[63531]: DEBUG nova.virt.hardware [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1101.558740] env[63531]: DEBUG nova.virt.hardware [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1101.559954] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de001391-b84e-4a55-8427-93d99cc8de33 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.571220] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69d5ee82-c89f-4743-8746-55d40de523cd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.627514] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23dd2a36-bd46-4513-825c-0e47d0e348d7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.638513] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "7fc38de1-1fbf-4312-98f6-6801d5fcd49a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1101.638930] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "7fc38de1-1fbf-4312-98f6-6801d5fcd49a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1101.645164] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e23a6dee-c74c-4a61-aa5c-a53648473303 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.690427] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28d288c1-49c6-4052-b9b7-1c32f528a260 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.698877] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51822c26-b3ea-4461-a71b-cc7a1ef0402a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.714790] env[63531]: DEBUG nova.compute.provider_tree [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Updating inventory in ProviderTree for provider 9996a314-a62e-4282-8252-f387e7e3fabb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1101.744919] env[63531]: DEBUG oslo_vmware.api [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]527623a6-bb90-ce4a-4f86-156fc33547f9, 'name': SearchDatastore_Task, 'duration_secs': 0.013717} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.750139] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Reconfiguring VM instance instance-00000063 to detach disk 2000 {{(pid=63531) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1101.750697] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-985e36b4-32ab-4765-a93a-5aa0028b52e0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.768575] env[63531]: DEBUG nova.compute.manager [req-4658fb57-029a-4c9f-801d-d6aac3315b79 req-3a4cfa8a-df47-4b5e-aa85-65f63ea2a9bb service nova] [instance: fbf63fa0-c8ec-41c7-878e-1446537b7b49] Received event network-vif-plugged-4fa0186c-186b-4f32-8db3-dc07235c8803 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1101.768804] env[63531]: DEBUG oslo_concurrency.lockutils [req-4658fb57-029a-4c9f-801d-d6aac3315b79 req-3a4cfa8a-df47-4b5e-aa85-65f63ea2a9bb service nova] Acquiring lock "fbf63fa0-c8ec-41c7-878e-1446537b7b49-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1101.769020] env[63531]: DEBUG oslo_concurrency.lockutils [req-4658fb57-029a-4c9f-801d-d6aac3315b79 req-3a4cfa8a-df47-4b5e-aa85-65f63ea2a9bb service nova] Lock "fbf63fa0-c8ec-41c7-878e-1446537b7b49-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1101.769193] env[63531]: DEBUG oslo_concurrency.lockutils [req-4658fb57-029a-4c9f-801d-d6aac3315b79 req-3a4cfa8a-df47-4b5e-aa85-65f63ea2a9bb service nova] Lock "fbf63fa0-c8ec-41c7-878e-1446537b7b49-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1101.769410] env[63531]: DEBUG nova.compute.manager [req-4658fb57-029a-4c9f-801d-d6aac3315b79 req-3a4cfa8a-df47-4b5e-aa85-65f63ea2a9bb service nova] [instance: fbf63fa0-c8ec-41c7-878e-1446537b7b49] No waiting events found dispatching network-vif-plugged-4fa0186c-186b-4f32-8db3-dc07235c8803 {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1101.769543] env[63531]: WARNING nova.compute.manager [req-4658fb57-029a-4c9f-801d-d6aac3315b79 req-3a4cfa8a-df47-4b5e-aa85-65f63ea2a9bb service nova] [instance: fbf63fa0-c8ec-41c7-878e-1446537b7b49] Received unexpected event network-vif-plugged-4fa0186c-186b-4f32-8db3-dc07235c8803 for instance with vm_state building and task_state spawning. [ 1101.771349] env[63531]: DEBUG oslo_vmware.api [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 1101.771349] env[63531]: value = "task-1118638" [ 1101.771349] env[63531]: _type = "Task" [ 1101.771349] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.780614] env[63531]: DEBUG oslo_vmware.api [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118638, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.907099] env[63531]: DEBUG nova.network.neutron [-] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1101.913195] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-050024a2-e6b9-4763-ad37-6f9e67ecc3f8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.935943] env[63531]: DEBUG oslo_vmware.api [None req-de477bf0-a47e-4909-bd9d-2464242d52bb tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118637, 'name': ReconfigVM_Task, 'duration_secs': 0.263428} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.937089] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-de477bf0-a47e-4909-bd9d-2464242d52bb tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Reconfigured VM instance instance-0000005c to detach disk 2001 {{(pid=63531) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1101.941875] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b9b7584d-78a6-4d9d-9d87-5e899a909dc5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.954090] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6739b050-98f8-4eaa-a5ba-5a21798fb7be {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.978153] env[63531]: DEBUG oslo_vmware.api [None req-de477bf0-a47e-4909-bd9d-2464242d52bb tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 1101.978153] env[63531]: value = "task-1118639" [ 1101.978153] env[63531]: _type = "Task" [ 1101.978153] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.996169] env[63531]: DEBUG nova.compute.manager [req-94d1f472-8d0b-4cff-8a0a-90ca8df0bb97 req-f03f6d49-0e06-48bf-8014-89f2c82e571e service nova] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] Detach interface failed, port_id=1765d3a3-d6ed-4150-aedf-45d90978133b, reason: Instance 148c6ecd-354e-4076-b52b-c13a2119da7e could not be found. {{(pid=63531) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1102.000077] env[63531]: INFO nova.compute.manager [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Took 16.77 seconds to build instance. [ 1102.004699] env[63531]: DEBUG oslo_vmware.api [None req-de477bf0-a47e-4909-bd9d-2464242d52bb tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118639, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.145068] env[63531]: DEBUG nova.compute.manager [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1102.249476] env[63531]: DEBUG nova.scheduler.client.report [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Updated inventory for provider 9996a314-a62e-4282-8252-f387e7e3fabb with generation 151 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1102.249629] env[63531]: DEBUG nova.compute.provider_tree [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Updating resource provider 9996a314-a62e-4282-8252-f387e7e3fabb generation from 151 to 152 during operation: update_inventory {{(pid=63531) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1102.249690] env[63531]: DEBUG nova.compute.provider_tree [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Updating inventory in ProviderTree for provider 9996a314-a62e-4282-8252-f387e7e3fabb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1102.282625] env[63531]: DEBUG oslo_vmware.api [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118638, 'name': ReconfigVM_Task, 'duration_secs': 0.360388} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.282903] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Reconfigured VM instance instance-00000063 to detach disk 2000 {{(pid=63531) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1102.283677] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d07892e8-0721-42eb-b94a-02502c24aa25 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.306184] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] 96c2910b-1e25-4053-9ad9-c155402ef0ac/96c2910b-1e25-4053-9ad9-c155402ef0ac.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1102.306475] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5ca263e6-cf25-4c41-9b3e-6662a8988dfc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.325064] env[63531]: DEBUG oslo_vmware.api [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 1102.325064] env[63531]: value = "task-1118640" [ 1102.325064] env[63531]: _type = "Task" [ 1102.325064] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.334834] env[63531]: DEBUG oslo_vmware.api [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118640, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.335698] env[63531]: DEBUG nova.network.neutron [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] [instance: fbf63fa0-c8ec-41c7-878e-1446537b7b49] Successfully updated port: 4fa0186c-186b-4f32-8db3-dc07235c8803 {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1102.409734] env[63531]: INFO nova.compute.manager [-] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] Took 1.30 seconds to deallocate network for instance. [ 1102.488778] env[63531]: DEBUG oslo_vmware.api [None req-de477bf0-a47e-4909-bd9d-2464242d52bb tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118639, 'name': ReconfigVM_Task, 'duration_secs': 0.152361} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.489164] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-de477bf0-a47e-4909-bd9d-2464242d52bb tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244792', 'volume_id': '5c5f7c58-05a8-42fc-9d14-8298686c8120', 'name': 'volume-5c5f7c58-05a8-42fc-9d14-8298686c8120', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': '18057005-1ff1-4e3d-bccc-8c89755d6db7', 'attached_at': '2024-10-03T08:25:29.000000', 'detached_at': '', 'volume_id': '5c5f7c58-05a8-42fc-9d14-8298686c8120', 'serial': '5c5f7c58-05a8-42fc-9d14-8298686c8120'} {{(pid=63531) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1102.489520] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-de477bf0-a47e-4909-bd9d-2464242d52bb tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1102.490324] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adc6a1f9-8c85-4ffb-9aca-9f56db320602 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.496964] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-de477bf0-a47e-4909-bd9d-2464242d52bb tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1102.497197] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-601bab2f-1852-4801-a633-d3918ccfb03e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.506077] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bd42c42e-6aa2-4c99-9863-0e6a5ba42bd7 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "ccc8f366-5849-45a6-9745-ff9b2be9c3bb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.289s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1102.570778] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-de477bf0-a47e-4909-bd9d-2464242d52bb tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1102.571094] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-de477bf0-a47e-4909-bd9d-2464242d52bb tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1102.571291] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-de477bf0-a47e-4909-bd9d-2464242d52bb tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Deleting the datastore file [datastore1] 18057005-1ff1-4e3d-bccc-8c89755d6db7 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1102.571942] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-74317128-9d61-4af5-8cae-00b1b09a5c86 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.579717] env[63531]: DEBUG oslo_vmware.api [None req-de477bf0-a47e-4909-bd9d-2464242d52bb tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 1102.579717] env[63531]: value = "task-1118642" [ 1102.579717] env[63531]: _type = "Task" [ 1102.579717] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.588213] env[63531]: DEBUG oslo_vmware.api [None req-de477bf0-a47e-4909-bd9d-2464242d52bb tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118642, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.669745] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1102.754244] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 3.262s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1102.754473] env[63531]: INFO nova.compute.manager [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Migrating [ 1102.761082] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.092s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1102.762475] env[63531]: INFO nova.compute.claims [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1102.834606] env[63531]: DEBUG oslo_vmware.api [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118640, 'name': ReconfigVM_Task, 'duration_secs': 0.49597} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.835041] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Reconfigured VM instance instance-00000063 to attach disk [datastore1] 96c2910b-1e25-4053-9ad9-c155402ef0ac/96c2910b-1e25-4053-9ad9-c155402ef0ac.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1102.836233] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d96f359-e660-4cb7-9ac8-512e76a229e1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.839308] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Acquiring lock "refresh_cache-fbf63fa0-c8ec-41c7-878e-1446537b7b49" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1102.839440] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Acquired lock "refresh_cache-fbf63fa0-c8ec-41c7-878e-1446537b7b49" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1102.839752] env[63531]: DEBUG nova.network.neutron [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] [instance: fbf63fa0-c8ec-41c7-878e-1446537b7b49] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1102.857729] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62395a2e-462b-4bdd-a6ec-1bd51636d230 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.878687] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6204da0c-5f23-4aa1-9ad4-88fe0ce537bc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.897873] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0deb20e-9213-40c8-8dc6-7594302972be {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.905109] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1102.905359] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4ce00190-a64a-4aa9-945a-04fc035e3547 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.911781] env[63531]: DEBUG oslo_vmware.api [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 1102.911781] env[63531]: value = "task-1118643" [ 1102.911781] env[63531]: _type = "Task" [ 1102.911781] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.915709] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c4d0011c-eb80-4408-9de7-9a716bf764ba tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1102.935063] env[63531]: DEBUG oslo_vmware.api [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118643, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.090854] env[63531]: DEBUG oslo_vmware.api [None req-de477bf0-a47e-4909-bd9d-2464242d52bb tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118642, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.17875} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.091176] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-de477bf0-a47e-4909-bd9d-2464242d52bb tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1103.091401] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-de477bf0-a47e-4909-bd9d-2464242d52bb tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1103.091637] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-de477bf0-a47e-4909-bd9d-2464242d52bb tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1103.091864] env[63531]: INFO nova.compute.manager [None req-de477bf0-a47e-4909-bd9d-2464242d52bb tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Took 2.31 seconds to destroy the instance on the hypervisor. [ 1103.092176] env[63531]: DEBUG oslo.service.loopingcall [None req-de477bf0-a47e-4909-bd9d-2464242d52bb tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1103.092408] env[63531]: DEBUG nova.compute.manager [-] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1103.092533] env[63531]: DEBUG nova.network.neutron [-] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1103.276323] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquiring lock "refresh_cache-49967454-a4f9-47f6-a2a9-0cd0c29e18d8" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1103.276323] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquired lock "refresh_cache-49967454-a4f9-47f6-a2a9-0cd0c29e18d8" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1103.276323] env[63531]: DEBUG nova.network.neutron [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1103.394893] env[63531]: DEBUG nova.network.neutron [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] [instance: fbf63fa0-c8ec-41c7-878e-1446537b7b49] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1103.425855] env[63531]: DEBUG oslo_vmware.api [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118643, 'name': PowerOnVM_Task, 'duration_secs': 0.463469} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.426348] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1103.443117] env[63531]: DEBUG nova.compute.manager [req-ba70e7bb-8656-4c6f-bfe4-5507265c9bde req-7d65f2d1-c392-4a23-8c56-a3ac8ca51139 service nova] [instance: fbf63fa0-c8ec-41c7-878e-1446537b7b49] Received event network-changed-4fa0186c-186b-4f32-8db3-dc07235c8803 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1103.443316] env[63531]: DEBUG nova.compute.manager [req-ba70e7bb-8656-4c6f-bfe4-5507265c9bde req-7d65f2d1-c392-4a23-8c56-a3ac8ca51139 service nova] [instance: fbf63fa0-c8ec-41c7-878e-1446537b7b49] Refreshing instance network info cache due to event network-changed-4fa0186c-186b-4f32-8db3-dc07235c8803. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1103.443661] env[63531]: DEBUG oslo_concurrency.lockutils [req-ba70e7bb-8656-4c6f-bfe4-5507265c9bde req-7d65f2d1-c392-4a23-8c56-a3ac8ca51139 service nova] Acquiring lock "refresh_cache-fbf63fa0-c8ec-41c7-878e-1446537b7b49" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1103.594587] env[63531]: DEBUG nova.network.neutron [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] [instance: fbf63fa0-c8ec-41c7-878e-1446537b7b49] Updating instance_info_cache with network_info: [{"id": "4fa0186c-186b-4f32-8db3-dc07235c8803", "address": "fa:16:3e:bc:1d:6c", "network": {"id": "423b0d4e-6f4f-4014-bdd3-f6f26e8233ed", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-2114093541-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "126f5465d9be40bf8b6d1dff02f10c7a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5a24f297-626b-4461-8c8d-1140fe436836", "external-id": "nsx-vlan-transportzone-604", "segmentation_id": 604, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4fa0186c-18", "ovs_interfaceid": "4fa0186c-186b-4f32-8db3-dc07235c8803", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1104.005903] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd906707-19d3-422a-83de-7549498fcaba {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.015720] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-219d872e-f81a-485a-9d40-6222b7405f94 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.050020] env[63531]: DEBUG nova.network.neutron [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Updating instance_info_cache with network_info: [{"id": "99a45e7c-1ddb-4c35-ac69-b2c33f976019", "address": "fa:16:3e:c8:19:0d", "network": {"id": "eac8786d-bc67-4916-8483-b53d9a25efb8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1604989690-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.241", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "abdae960d43b47bc95ab41f265d2c85b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db068f71-08cc-42d4-8ab6-17134c1585e5", "external-id": "nsx-vlan-transportzone-721", "segmentation_id": 721, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap99a45e7c-1d", "ovs_interfaceid": "99a45e7c-1ddb-4c35-ac69-b2c33f976019", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1104.051765] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-667ac36d-810a-4a0e-a9ee-90b77a862183 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.062599] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdc4b0db-8094-4734-96d5-61fc863b4d80 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.077914] env[63531]: DEBUG nova.compute.provider_tree [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Updating inventory in ProviderTree for provider 9996a314-a62e-4282-8252-f387e7e3fabb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1104.099770] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Releasing lock "refresh_cache-fbf63fa0-c8ec-41c7-878e-1446537b7b49" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1104.100152] env[63531]: DEBUG nova.compute.manager [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] [instance: fbf63fa0-c8ec-41c7-878e-1446537b7b49] Instance network_info: |[{"id": "4fa0186c-186b-4f32-8db3-dc07235c8803", "address": "fa:16:3e:bc:1d:6c", "network": {"id": "423b0d4e-6f4f-4014-bdd3-f6f26e8233ed", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-2114093541-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "126f5465d9be40bf8b6d1dff02f10c7a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5a24f297-626b-4461-8c8d-1140fe436836", "external-id": "nsx-vlan-transportzone-604", "segmentation_id": 604, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4fa0186c-18", "ovs_interfaceid": "4fa0186c-186b-4f32-8db3-dc07235c8803", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1104.100434] env[63531]: DEBUG oslo_concurrency.lockutils [req-ba70e7bb-8656-4c6f-bfe4-5507265c9bde req-7d65f2d1-c392-4a23-8c56-a3ac8ca51139 service nova] Acquired lock "refresh_cache-fbf63fa0-c8ec-41c7-878e-1446537b7b49" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1104.100594] env[63531]: DEBUG nova.network.neutron [req-ba70e7bb-8656-4c6f-bfe4-5507265c9bde req-7d65f2d1-c392-4a23-8c56-a3ac8ca51139 service nova] [instance: fbf63fa0-c8ec-41c7-878e-1446537b7b49] Refreshing network info cache for port 4fa0186c-186b-4f32-8db3-dc07235c8803 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1104.101820] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] [instance: fbf63fa0-c8ec-41c7-878e-1446537b7b49] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bc:1d:6c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5a24f297-626b-4461-8c8d-1140fe436836', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4fa0186c-186b-4f32-8db3-dc07235c8803', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1104.109012] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Creating folder: Project (126f5465d9be40bf8b6d1dff02f10c7a). Parent ref: group-v244585. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1104.111921] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4fa9154a-f7f6-4692-9191-ace3475de645 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.124706] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Created folder: Project (126f5465d9be40bf8b6d1dff02f10c7a) in parent group-v244585. [ 1104.124927] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Creating folder: Instances. Parent ref: group-v244807. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1104.125195] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-15a79104-f829-466f-ba18-635792e1e8e9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.136053] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Created folder: Instances in parent group-v244807. [ 1104.136341] env[63531]: DEBUG oslo.service.loopingcall [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1104.136777] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fbf63fa0-c8ec-41c7-878e-1446537b7b49] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1104.136777] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2e16bd85-a6bf-4447-a77c-b2f09df45e36 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.158882] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1104.158882] env[63531]: value = "task-1118646" [ 1104.158882] env[63531]: _type = "Task" [ 1104.158882] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.168036] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118646, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.212525] env[63531]: DEBUG nova.network.neutron [-] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1104.220869] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "98853767-791e-4568-8bb8-30dfcc473ee3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1104.221166] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "98853767-791e-4568-8bb8-30dfcc473ee3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1104.340019] env[63531]: DEBUG nova.network.neutron [req-ba70e7bb-8656-4c6f-bfe4-5507265c9bde req-7d65f2d1-c392-4a23-8c56-a3ac8ca51139 service nova] [instance: fbf63fa0-c8ec-41c7-878e-1446537b7b49] Updated VIF entry in instance network info cache for port 4fa0186c-186b-4f32-8db3-dc07235c8803. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1104.340427] env[63531]: DEBUG nova.network.neutron [req-ba70e7bb-8656-4c6f-bfe4-5507265c9bde req-7d65f2d1-c392-4a23-8c56-a3ac8ca51139 service nova] [instance: fbf63fa0-c8ec-41c7-878e-1446537b7b49] Updating instance_info_cache with network_info: [{"id": "4fa0186c-186b-4f32-8db3-dc07235c8803", "address": "fa:16:3e:bc:1d:6c", "network": {"id": "423b0d4e-6f4f-4014-bdd3-f6f26e8233ed", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-2114093541-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "126f5465d9be40bf8b6d1dff02f10c7a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5a24f297-626b-4461-8c8d-1140fe436836", "external-id": "nsx-vlan-transportzone-604", "segmentation_id": 604, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4fa0186c-18", "ovs_interfaceid": "4fa0186c-186b-4f32-8db3-dc07235c8803", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1104.439811] env[63531]: INFO nova.compute.manager [None req-2928b464-5e45-495d-8a0f-869a2d137447 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Updating instance to original state: 'active' [ 1104.556730] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Releasing lock "refresh_cache-49967454-a4f9-47f6-a2a9-0cd0c29e18d8" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1104.599864] env[63531]: ERROR nova.scheduler.client.report [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [req-4648f914-6584-4acc-92d7-4692841c8f21] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 9996a314-a62e-4282-8252-f387e7e3fabb. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-4648f914-6584-4acc-92d7-4692841c8f21"}]} [ 1104.615331] env[63531]: DEBUG nova.scheduler.client.report [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Refreshing inventories for resource provider 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1104.629405] env[63531]: DEBUG nova.scheduler.client.report [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Updating ProviderTree inventory for provider 9996a314-a62e-4282-8252-f387e7e3fabb from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1104.629729] env[63531]: DEBUG nova.compute.provider_tree [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Updating inventory in ProviderTree for provider 9996a314-a62e-4282-8252-f387e7e3fabb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1104.639448] env[63531]: DEBUG nova.scheduler.client.report [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Refreshing aggregate associations for resource provider 9996a314-a62e-4282-8252-f387e7e3fabb, aggregates: None {{(pid=63531) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1104.655659] env[63531]: DEBUG nova.scheduler.client.report [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Refreshing trait associations for resource provider 9996a314-a62e-4282-8252-f387e7e3fabb, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE {{(pid=63531) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1104.668757] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118646, 'name': CreateVM_Task} progress is 25%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.717963] env[63531]: INFO nova.compute.manager [-] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Took 1.63 seconds to deallocate network for instance. [ 1104.725899] env[63531]: DEBUG nova.compute.manager [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1104.779045] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1104.779611] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1104.779968] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Starting heal instance info cache {{(pid=63531) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1104.840669] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-276d1807-f952-4dbf-bb79-89c54242db27 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.845151] env[63531]: DEBUG oslo_concurrency.lockutils [req-ba70e7bb-8656-4c6f-bfe4-5507265c9bde req-7d65f2d1-c392-4a23-8c56-a3ac8ca51139 service nova] Releasing lock "refresh_cache-fbf63fa0-c8ec-41c7-878e-1446537b7b49" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1104.845151] env[63531]: DEBUG nova.compute.manager [req-ba70e7bb-8656-4c6f-bfe4-5507265c9bde req-7d65f2d1-c392-4a23-8c56-a3ac8ca51139 service nova] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Received event network-changed-e9692fb2-f3a9-455c-8317-b533a6af48f7 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1104.845151] env[63531]: DEBUG nova.compute.manager [req-ba70e7bb-8656-4c6f-bfe4-5507265c9bde req-7d65f2d1-c392-4a23-8c56-a3ac8ca51139 service nova] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Refreshing instance network info cache due to event network-changed-e9692fb2-f3a9-455c-8317-b533a6af48f7. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1104.845151] env[63531]: DEBUG oslo_concurrency.lockutils [req-ba70e7bb-8656-4c6f-bfe4-5507265c9bde req-7d65f2d1-c392-4a23-8c56-a3ac8ca51139 service nova] Acquiring lock "refresh_cache-ccc8f366-5849-45a6-9745-ff9b2be9c3bb" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1104.845151] env[63531]: DEBUG oslo_concurrency.lockutils [req-ba70e7bb-8656-4c6f-bfe4-5507265c9bde req-7d65f2d1-c392-4a23-8c56-a3ac8ca51139 service nova] Acquired lock "refresh_cache-ccc8f366-5849-45a6-9745-ff9b2be9c3bb" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1104.845151] env[63531]: DEBUG nova.network.neutron [req-ba70e7bb-8656-4c6f-bfe4-5507265c9bde req-7d65f2d1-c392-4a23-8c56-a3ac8ca51139 service nova] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Refreshing network info cache for port e9692fb2-f3a9-455c-8317-b533a6af48f7 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1104.851460] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d7b2648-94d7-4e63-b537-cab48943e406 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.883339] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-950ed1bd-f4a9-4b91-bf0b-b10b5335dd80 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.891244] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9943d99a-7822-477f-9704-db7e141a555b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.904758] env[63531]: DEBUG nova.compute.provider_tree [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Updating inventory in ProviderTree for provider 9996a314-a62e-4282-8252-f387e7e3fabb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1105.169988] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118646, 'name': CreateVM_Task, 'duration_secs': 0.702857} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.170175] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fbf63fa0-c8ec-41c7-878e-1446537b7b49] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1105.170895] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1105.171085] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1105.171413] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1105.171656] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0316f60b-dfab-43d1-b53b-b7c09bbe60df {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.177851] env[63531]: DEBUG oslo_vmware.api [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Waiting for the task: (returnval){ [ 1105.177851] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]525111b3-e41a-47e1-afb1-472ddc669841" [ 1105.177851] env[63531]: _type = "Task" [ 1105.177851] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.185169] env[63531]: DEBUG oslo_vmware.api [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]525111b3-e41a-47e1-afb1-472ddc669841, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.245558] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1105.261675] env[63531]: INFO nova.compute.manager [None req-de477bf0-a47e-4909-bd9d-2464242d52bb tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Took 0.54 seconds to detach 1 volumes for instance. [ 1105.314449] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Acquiring lock "refresh_cache-5ad9fc53-104b-4a90-801a-bd1e8f12c1d7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1105.314609] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Acquired lock "refresh_cache-5ad9fc53-104b-4a90-801a-bd1e8f12c1d7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1105.314760] env[63531]: DEBUG nova.network.neutron [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] Forcefully refreshing network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1105.435712] env[63531]: DEBUG nova.scheduler.client.report [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Updated inventory for provider 9996a314-a62e-4282-8252-f387e7e3fabb with generation 153 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1105.436017] env[63531]: DEBUG nova.compute.provider_tree [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Updating resource provider 9996a314-a62e-4282-8252-f387e7e3fabb generation from 153 to 154 during operation: update_inventory {{(pid=63531) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1105.436189] env[63531]: DEBUG nova.compute.provider_tree [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Updating inventory in ProviderTree for provider 9996a314-a62e-4282-8252-f387e7e3fabb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1105.477294] env[63531]: DEBUG nova.compute.manager [req-4a11a157-5c25-4da4-9f10-1ad606f2b920 req-bf1f4963-5b39-4451-8fee-5b0a195b8f63 service nova] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Received event network-vif-deleted-af2485f6-5ce2-4a83-9d95-7bea85cebff2 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1105.574179] env[63531]: DEBUG nova.network.neutron [req-ba70e7bb-8656-4c6f-bfe4-5507265c9bde req-7d65f2d1-c392-4a23-8c56-a3ac8ca51139 service nova] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Updated VIF entry in instance network info cache for port e9692fb2-f3a9-455c-8317-b533a6af48f7. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1105.574571] env[63531]: DEBUG nova.network.neutron [req-ba70e7bb-8656-4c6f-bfe4-5507265c9bde req-7d65f2d1-c392-4a23-8c56-a3ac8ca51139 service nova] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Updating instance_info_cache with network_info: [{"id": "e9692fb2-f3a9-455c-8317-b533a6af48f7", "address": "fa:16:3e:79:4a:3b", "network": {"id": "caf86428-59a5-4e50-9135-2ab8473af64d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1673994696-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ba8b284854242c392aec5326e996239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape9692fb2-f3", "ovs_interfaceid": "e9692fb2-f3a9-455c-8317-b533a6af48f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1105.687696] env[63531]: DEBUG oslo_vmware.api [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]525111b3-e41a-47e1-afb1-472ddc669841, 'name': SearchDatastore_Task, 'duration_secs': 0.010704} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.688057] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1105.688307] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] [instance: fbf63fa0-c8ec-41c7-878e-1446537b7b49] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1105.688552] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1105.688961] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1105.689206] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1105.689451] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e7a1ff2f-0a54-4d60-91f0-db6cfccf568c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.700286] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1105.700672] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1105.701441] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2f204e3d-1278-40cb-85a7-c0c2113a4bc4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.707408] env[63531]: DEBUG oslo_vmware.api [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Waiting for the task: (returnval){ [ 1105.707408] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52eeabe6-991e-6bcf-cba7-fc8fb77d7874" [ 1105.707408] env[63531]: _type = "Task" [ 1105.707408] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.717088] env[63531]: DEBUG oslo_vmware.api [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52eeabe6-991e-6bcf-cba7-fc8fb77d7874, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.770562] env[63531]: DEBUG oslo_concurrency.lockutils [None req-de477bf0-a47e-4909-bd9d-2464242d52bb tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1105.941394] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.180s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1105.942021] env[63531]: DEBUG nova.compute.manager [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1105.944490] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c4d0011c-eb80-4408-9de7-9a716bf764ba tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.029s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1105.944727] env[63531]: DEBUG nova.objects.instance [None req-c4d0011c-eb80-4408-9de7-9a716bf764ba tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lazy-loading 'resources' on Instance uuid 148c6ecd-354e-4076-b52b-c13a2119da7e {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1106.071897] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f82f0bc-2565-4ddc-b16e-ec7876100b61 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.076593] env[63531]: DEBUG oslo_concurrency.lockutils [req-ba70e7bb-8656-4c6f-bfe4-5507265c9bde req-7d65f2d1-c392-4a23-8c56-a3ac8ca51139 service nova] Releasing lock "refresh_cache-ccc8f366-5849-45a6-9745-ff9b2be9c3bb" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1106.095128] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Updating instance '49967454-a4f9-47f6-a2a9-0cd0c29e18d8' progress to 0 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1106.218449] env[63531]: DEBUG oslo_vmware.api [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52eeabe6-991e-6bcf-cba7-fc8fb77d7874, 'name': SearchDatastore_Task, 'duration_secs': 0.012934} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.219242] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e8b3c66c-880b-4815-b4c4-7cf4c41315a8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.224726] env[63531]: DEBUG oslo_vmware.api [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Waiting for the task: (returnval){ [ 1106.224726] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52a0d945-7e38-1df3-9542-9019f9c7c748" [ 1106.224726] env[63531]: _type = "Task" [ 1106.224726] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.232024] env[63531]: DEBUG oslo_vmware.api [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52a0d945-7e38-1df3-9542-9019f9c7c748, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.413918] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ca929e81-526c-4b49-8799-8ac6635bcbc4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "96c2910b-1e25-4053-9ad9-c155402ef0ac" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1106.414198] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ca929e81-526c-4b49-8799-8ac6635bcbc4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "96c2910b-1e25-4053-9ad9-c155402ef0ac" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1106.414416] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ca929e81-526c-4b49-8799-8ac6635bcbc4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "96c2910b-1e25-4053-9ad9-c155402ef0ac-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1106.414600] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ca929e81-526c-4b49-8799-8ac6635bcbc4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "96c2910b-1e25-4053-9ad9-c155402ef0ac-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1106.414776] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ca929e81-526c-4b49-8799-8ac6635bcbc4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "96c2910b-1e25-4053-9ad9-c155402ef0ac-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1106.416735] env[63531]: INFO nova.compute.manager [None req-ca929e81-526c-4b49-8799-8ac6635bcbc4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Terminating instance [ 1106.420484] env[63531]: DEBUG nova.compute.manager [None req-ca929e81-526c-4b49-8799-8ac6635bcbc4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1106.420685] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-ca929e81-526c-4b49-8799-8ac6635bcbc4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1106.421845] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da46bd62-402c-4c82-9cbd-7303bd116bd8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.430391] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca929e81-526c-4b49-8799-8ac6635bcbc4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1106.430620] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8f5cd526-0ce3-4c9a-a3f4-541a5d889614 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.437755] env[63531]: DEBUG oslo_vmware.api [None req-ca929e81-526c-4b49-8799-8ac6635bcbc4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 1106.437755] env[63531]: value = "task-1118647" [ 1106.437755] env[63531]: _type = "Task" [ 1106.437755] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.445421] env[63531]: DEBUG oslo_vmware.api [None req-ca929e81-526c-4b49-8799-8ac6635bcbc4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118647, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.449015] env[63531]: DEBUG nova.compute.utils [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1106.452446] env[63531]: DEBUG nova.compute.manager [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1106.452616] env[63531]: DEBUG nova.network.neutron [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1106.518356] env[63531]: DEBUG nova.policy [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7140e12106504b3e95ff49786361771b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '63900312a30b4942854f552758c039cf', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 1106.593777] env[63531]: DEBUG nova.network.neutron [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] Updating instance_info_cache with network_info: [{"id": "161e87a1-1133-476f-9598-44c29595c36a", "address": "fa:16:3e:f3:d5:51", "network": {"id": "80641cc0-680b-40ff-8d1b-f5a5f736a88f", "bridge": "br-int", "label": "tempest-ServersTestJSON-666288463-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "63900312a30b4942854f552758c039cf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "26472e27-9835-4f87-ab7f-ca24dfee4e83", "external-id": "nsx-vlan-transportzone-335", "segmentation_id": 335, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap161e87a1-11", "ovs_interfaceid": "161e87a1-1133-476f-9598-44c29595c36a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1106.601391] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1106.601949] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a877b9c3-8c59-4875-89c8-81279879d28a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.610978] env[63531]: DEBUG oslo_vmware.api [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 1106.610978] env[63531]: value = "task-1118648" [ 1106.610978] env[63531]: _type = "Task" [ 1106.610978] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.624863] env[63531]: DEBUG oslo_vmware.api [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118648, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.631419] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-540faed3-fd14-45ea-b2b7-802935bb5780 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.639976] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-457f27d8-6d11-4f2e-b5fb-8d4e2b8014b1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.674517] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7517ccf-1d7f-400d-af57-ecdf46e6b1c6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.682713] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce6eb70d-841a-474c-9bc2-502f8679dd6a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.697655] env[63531]: DEBUG nova.compute.provider_tree [None req-c4d0011c-eb80-4408-9de7-9a716bf764ba tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1106.735622] env[63531]: DEBUG oslo_vmware.api [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52a0d945-7e38-1df3-9542-9019f9c7c748, 'name': SearchDatastore_Task, 'duration_secs': 0.039828} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.736024] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1106.736276] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] fbf63fa0-c8ec-41c7-878e-1446537b7b49/fbf63fa0-c8ec-41c7-878e-1446537b7b49.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1106.736566] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-02978008-ebc6-43df-becb-f1dcee516279 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.744462] env[63531]: DEBUG oslo_vmware.api [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Waiting for the task: (returnval){ [ 1106.744462] env[63531]: value = "task-1118649" [ 1106.744462] env[63531]: _type = "Task" [ 1106.744462] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.753863] env[63531]: DEBUG oslo_vmware.api [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Task: {'id': task-1118649, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.847147] env[63531]: DEBUG nova.network.neutron [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] Successfully created port: e22c7894-d0a9-4984-993f-0c220a12dfd5 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1106.947695] env[63531]: DEBUG oslo_vmware.api [None req-ca929e81-526c-4b49-8799-8ac6635bcbc4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118647, 'name': PowerOffVM_Task, 'duration_secs': 0.355358} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.948016] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca929e81-526c-4b49-8799-8ac6635bcbc4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1106.948204] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-ca929e81-526c-4b49-8799-8ac6635bcbc4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1106.948473] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-67022d0a-4cc9-4c6a-909f-775a08f3a2b9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.955794] env[63531]: DEBUG nova.compute.manager [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1107.029188] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-ca929e81-526c-4b49-8799-8ac6635bcbc4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1107.029452] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-ca929e81-526c-4b49-8799-8ac6635bcbc4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1107.029633] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca929e81-526c-4b49-8799-8ac6635bcbc4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Deleting the datastore file [datastore1] 96c2910b-1e25-4053-9ad9-c155402ef0ac {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1107.029946] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-eb2ec00e-2ce4-4b29-9600-4eaa26568736 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.038091] env[63531]: DEBUG oslo_vmware.api [None req-ca929e81-526c-4b49-8799-8ac6635bcbc4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 1107.038091] env[63531]: value = "task-1118651" [ 1107.038091] env[63531]: _type = "Task" [ 1107.038091] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.046153] env[63531]: DEBUG oslo_vmware.api [None req-ca929e81-526c-4b49-8799-8ac6635bcbc4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118651, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.096501] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Releasing lock "refresh_cache-5ad9fc53-104b-4a90-801a-bd1e8f12c1d7" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1107.096727] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] Updated the network info_cache for instance {{(pid=63531) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1107.096949] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1107.097137] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1107.097294] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1107.097436] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1107.097589] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1107.097816] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63531) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1107.098037] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager.update_available_resource {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1107.122935] env[63531]: DEBUG oslo_vmware.api [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118648, 'name': PowerOffVM_Task, 'duration_secs': 0.195909} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.123339] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1107.123547] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Updating instance '49967454-a4f9-47f6-a2a9-0cd0c29e18d8' progress to 17 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1107.201446] env[63531]: DEBUG nova.scheduler.client.report [None req-c4d0011c-eb80-4408-9de7-9a716bf764ba tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1107.255326] env[63531]: DEBUG oslo_vmware.api [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Task: {'id': task-1118649, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.549703] env[63531]: DEBUG oslo_vmware.api [None req-ca929e81-526c-4b49-8799-8ac6635bcbc4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118651, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.439704} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.550066] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca929e81-526c-4b49-8799-8ac6635bcbc4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1107.550222] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-ca929e81-526c-4b49-8799-8ac6635bcbc4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1107.550390] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-ca929e81-526c-4b49-8799-8ac6635bcbc4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1107.550567] env[63531]: INFO nova.compute.manager [None req-ca929e81-526c-4b49-8799-8ac6635bcbc4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1107.550839] env[63531]: DEBUG oslo.service.loopingcall [None req-ca929e81-526c-4b49-8799-8ac6635bcbc4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1107.551063] env[63531]: DEBUG nova.compute.manager [-] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1107.551218] env[63531]: DEBUG nova.network.neutron [-] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1107.601694] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1107.630319] env[63531]: DEBUG nova.virt.hardware [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1107.630566] env[63531]: DEBUG nova.virt.hardware [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1107.630724] env[63531]: DEBUG nova.virt.hardware [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1107.631065] env[63531]: DEBUG nova.virt.hardware [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1107.631339] env[63531]: DEBUG nova.virt.hardware [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1107.631514] env[63531]: DEBUG nova.virt.hardware [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1107.631734] env[63531]: DEBUG nova.virt.hardware [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1107.631925] env[63531]: DEBUG nova.virt.hardware [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1107.632156] env[63531]: DEBUG nova.virt.hardware [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1107.632333] env[63531]: DEBUG nova.virt.hardware [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1107.632515] env[63531]: DEBUG nova.virt.hardware [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1107.637385] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cf5fa950-1dff-4277-9c90-cc98feb5c8d5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.658361] env[63531]: DEBUG oslo_vmware.api [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 1107.658361] env[63531]: value = "task-1118652" [ 1107.658361] env[63531]: _type = "Task" [ 1107.658361] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.667114] env[63531]: DEBUG oslo_vmware.api [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118652, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.706253] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c4d0011c-eb80-4408-9de7-9a716bf764ba tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.762s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1107.708852] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.463s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1107.710444] env[63531]: INFO nova.compute.claims [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1107.733533] env[63531]: INFO nova.scheduler.client.report [None req-c4d0011c-eb80-4408-9de7-9a716bf764ba tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Deleted allocations for instance 148c6ecd-354e-4076-b52b-c13a2119da7e [ 1107.755362] env[63531]: DEBUG oslo_vmware.api [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Task: {'id': task-1118649, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.675161} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.755642] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] fbf63fa0-c8ec-41c7-878e-1446537b7b49/fbf63fa0-c8ec-41c7-878e-1446537b7b49.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1107.755873] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] [instance: fbf63fa0-c8ec-41c7-878e-1446537b7b49] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1107.756308] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-22096e5e-1b8f-4381-b22b-e321a3438bb4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.764494] env[63531]: DEBUG oslo_vmware.api [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Waiting for the task: (returnval){ [ 1107.764494] env[63531]: value = "task-1118653" [ 1107.764494] env[63531]: _type = "Task" [ 1107.764494] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.777074] env[63531]: DEBUG oslo_vmware.api [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Task: {'id': task-1118653, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.967732] env[63531]: DEBUG nova.compute.manager [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1107.999289] env[63531]: DEBUG nova.virt.hardware [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1107.999558] env[63531]: DEBUG nova.virt.hardware [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1107.999728] env[63531]: DEBUG nova.virt.hardware [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1107.999945] env[63531]: DEBUG nova.virt.hardware [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1108.000126] env[63531]: DEBUG nova.virt.hardware [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1108.000428] env[63531]: DEBUG nova.virt.hardware [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1108.000509] env[63531]: DEBUG nova.virt.hardware [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1108.000645] env[63531]: DEBUG nova.virt.hardware [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1108.000816] env[63531]: DEBUG nova.virt.hardware [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1108.001055] env[63531]: DEBUG nova.virt.hardware [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1108.001222] env[63531]: DEBUG nova.virt.hardware [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1108.002075] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdde4663-9654-4d6f-9cec-401b82136e23 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.013281] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f055f65a-4dab-4940-8444-eedb4b025e05 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.021277] env[63531]: DEBUG nova.compute.manager [req-3c9af043-ff33-4c06-86d9-33592e3e614b req-f3f3f136-4e52-4c31-9b56-51f2a2a6b9e7 service nova] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Received event network-vif-deleted-b731766f-738c-4d77-a00e-98ed6c683e55 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1108.021609] env[63531]: INFO nova.compute.manager [req-3c9af043-ff33-4c06-86d9-33592e3e614b req-f3f3f136-4e52-4c31-9b56-51f2a2a6b9e7 service nova] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Neutron deleted interface b731766f-738c-4d77-a00e-98ed6c683e55; detaching it from the instance and deleting it from the info cache [ 1108.021911] env[63531]: DEBUG nova.network.neutron [req-3c9af043-ff33-4c06-86d9-33592e3e614b req-f3f3f136-4e52-4c31-9b56-51f2a2a6b9e7 service nova] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1108.171984] env[63531]: DEBUG oslo_vmware.api [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118652, 'name': ReconfigVM_Task, 'duration_secs': 0.195631} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.172339] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Updating instance '49967454-a4f9-47f6-a2a9-0cd0c29e18d8' progress to 33 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1108.242388] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c4d0011c-eb80-4408-9de7-9a716bf764ba tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "148c6ecd-354e-4076-b52b-c13a2119da7e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.766s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1108.274958] env[63531]: DEBUG oslo_vmware.api [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Task: {'id': task-1118653, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.086209} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.275303] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] [instance: fbf63fa0-c8ec-41c7-878e-1446537b7b49] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1108.276118] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14e104f6-5bad-4426-ab92-5463c8e3b471 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.299759] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] [instance: fbf63fa0-c8ec-41c7-878e-1446537b7b49] Reconfiguring VM instance instance-0000006e to attach disk [datastore1] fbf63fa0-c8ec-41c7-878e-1446537b7b49/fbf63fa0-c8ec-41c7-878e-1446537b7b49.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1108.300385] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-49822108-accb-4709-9f89-20ab12fa41ba {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.321611] env[63531]: DEBUG oslo_vmware.api [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Waiting for the task: (returnval){ [ 1108.321611] env[63531]: value = "task-1118654" [ 1108.321611] env[63531]: _type = "Task" [ 1108.321611] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.331192] env[63531]: DEBUG oslo_vmware.api [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Task: {'id': task-1118654, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.333132] env[63531]: DEBUG nova.network.neutron [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] Successfully updated port: e22c7894-d0a9-4984-993f-0c220a12dfd5 {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1108.495514] env[63531]: DEBUG nova.network.neutron [-] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1108.533948] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-277c91f7-26cf-4e8d-a51d-fc1f4cc731a4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.543626] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8625584-6bf9-436a-a811-7d8f7351d096 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.575820] env[63531]: DEBUG nova.compute.manager [req-3c9af043-ff33-4c06-86d9-33592e3e614b req-f3f3f136-4e52-4c31-9b56-51f2a2a6b9e7 service nova] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Detach interface failed, port_id=b731766f-738c-4d77-a00e-98ed6c683e55, reason: Instance 96c2910b-1e25-4053-9ad9-c155402ef0ac could not be found. {{(pid=63531) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1108.679616] env[63531]: DEBUG nova.virt.hardware [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1108.679616] env[63531]: DEBUG nova.virt.hardware [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1108.679864] env[63531]: DEBUG nova.virt.hardware [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1108.679864] env[63531]: DEBUG nova.virt.hardware [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1108.679992] env[63531]: DEBUG nova.virt.hardware [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1108.680176] env[63531]: DEBUG nova.virt.hardware [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1108.680369] env[63531]: DEBUG nova.virt.hardware [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1108.680530] env[63531]: DEBUG nova.virt.hardware [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1108.681725] env[63531]: DEBUG nova.virt.hardware [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1108.681725] env[63531]: DEBUG nova.virt.hardware [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1108.681725] env[63531]: DEBUG nova.virt.hardware [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1108.833619] env[63531]: DEBUG oslo_vmware.api [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Task: {'id': task-1118654, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.836892] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "refresh_cache-7fc38de1-1fbf-4312-98f6-6801d5fcd49a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1108.836892] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquired lock "refresh_cache-7fc38de1-1fbf-4312-98f6-6801d5fcd49a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1108.836892] env[63531]: DEBUG nova.network.neutron [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1108.998420] env[63531]: INFO nova.compute.manager [-] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Took 1.45 seconds to deallocate network for instance. [ 1109.033535] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cc3b7d6-e248-4675-9cda-c01ba7813102 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.041934] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c862b4c9-f740-4a8e-a037-3993a36316a7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.077493] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-654361b6-4252-4198-85ad-330d4cec1c5f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.085731] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-788d1c74-7c75-43cd-8170-4d7ed1424f16 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.100185] env[63531]: DEBUG nova.compute.provider_tree [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1109.189942] env[63531]: ERROR nova.compute.manager [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Setting instance vm_state to ERROR: AttributeError: 'NoneType' object has no attribute 'key' [ 1109.189942] env[63531]: ERROR nova.compute.manager [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Traceback (most recent call last): [ 1109.189942] env[63531]: ERROR nova.compute.manager [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] File "/opt/stack/nova/nova/compute/manager.py", line 10865, in _error_out_instance_on_exception [ 1109.189942] env[63531]: ERROR nova.compute.manager [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] yield [ 1109.189942] env[63531]: ERROR nova.compute.manager [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] File "/opt/stack/nova/nova/compute/manager.py", line 6105, in _resize_instance [ 1109.189942] env[63531]: ERROR nova.compute.manager [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] disk_info = self.driver.migrate_disk_and_power_off( [ 1109.189942] env[63531]: ERROR nova.compute.manager [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 264, in migrate_disk_and_power_off [ 1109.189942] env[63531]: ERROR nova.compute.manager [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] return self._vmops.migrate_disk_and_power_off(context, instance, [ 1109.189942] env[63531]: ERROR nova.compute.manager [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1467, in migrate_disk_and_power_off [ 1109.189942] env[63531]: ERROR nova.compute.manager [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] self._resize_disk(instance, vm_ref, vmdk, flavor) [ 1109.189942] env[63531]: ERROR nova.compute.manager [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1398, in _resize_disk [ 1109.189942] env[63531]: ERROR nova.compute.manager [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] self._volumeops.detach_disk_from_vm(vm_ref, instance, vmdk.device) [ 1109.189942] env[63531]: ERROR nova.compute.manager [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 121, in detach_disk_from_vm [ 1109.189942] env[63531]: ERROR nova.compute.manager [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] disk_key = device.key [ 1109.189942] env[63531]: ERROR nova.compute.manager [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] AttributeError: 'NoneType' object has no attribute 'key' [ 1109.189942] env[63531]: ERROR nova.compute.manager [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] [ 1109.336401] env[63531]: DEBUG oslo_vmware.api [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Task: {'id': task-1118654, 'name': ReconfigVM_Task, 'duration_secs': 0.738917} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.336728] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] [instance: fbf63fa0-c8ec-41c7-878e-1446537b7b49] Reconfigured VM instance instance-0000006e to attach disk [datastore1] fbf63fa0-c8ec-41c7-878e-1446537b7b49/fbf63fa0-c8ec-41c7-878e-1446537b7b49.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1109.337380] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-153ecb31-cebe-4d55-aba7-e871797e57f2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.347888] env[63531]: DEBUG oslo_vmware.api [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Waiting for the task: (returnval){ [ 1109.347888] env[63531]: value = "task-1118655" [ 1109.347888] env[63531]: _type = "Task" [ 1109.347888] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.356835] env[63531]: DEBUG oslo_vmware.api [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Task: {'id': task-1118655, 'name': Rename_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.366984] env[63531]: DEBUG nova.network.neutron [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1109.501362] env[63531]: DEBUG nova.network.neutron [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] Updating instance_info_cache with network_info: [{"id": "e22c7894-d0a9-4984-993f-0c220a12dfd5", "address": "fa:16:3e:84:57:e0", "network": {"id": "80641cc0-680b-40ff-8d1b-f5a5f736a88f", "bridge": "br-int", "label": "tempest-ServersTestJSON-666288463-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "63900312a30b4942854f552758c039cf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "26472e27-9835-4f87-ab7f-ca24dfee4e83", "external-id": "nsx-vlan-transportzone-335", "segmentation_id": 335, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape22c7894-d0", "ovs_interfaceid": "e22c7894-d0a9-4984-993f-0c220a12dfd5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1109.507885] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ca929e81-526c-4b49-8799-8ac6635bcbc4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1109.604060] env[63531]: DEBUG nova.scheduler.client.report [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1109.707403] env[63531]: INFO nova.compute.manager [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Swapping old allocation on dict_keys(['9996a314-a62e-4282-8252-f387e7e3fabb']) held by migration 72e41fa6-b1eb-48d8-9701-d5fc7cd368c7 for instance [ 1109.728123] env[63531]: DEBUG nova.scheduler.client.report [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Overwriting current allocation {'allocations': {'9996a314-a62e-4282-8252-f387e7e3fabb': {'resources': {'VCPU': 1, 'MEMORY_MB': 256}, 'generation': 155}}, 'project_id': 'abdae960d43b47bc95ab41f265d2c85b', 'user_id': '0a501b2c64c04dcda3596c8a1bd3fa0d', 'consumer_generation': 1} on consumer 49967454-a4f9-47f6-a2a9-0cd0c29e18d8 {{(pid=63531) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1109.857620] env[63531]: DEBUG oslo_vmware.api [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Task: {'id': task-1118655, 'name': Rename_Task, 'duration_secs': 0.146828} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.857906] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] [instance: fbf63fa0-c8ec-41c7-878e-1446537b7b49] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1109.858174] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1a5828af-b3ed-44fa-b4b7-f32f71f9dc22 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.865216] env[63531]: DEBUG oslo_vmware.api [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Waiting for the task: (returnval){ [ 1109.865216] env[63531]: value = "task-1118656" [ 1109.865216] env[63531]: _type = "Task" [ 1109.865216] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.873097] env[63531]: DEBUG oslo_vmware.api [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Task: {'id': task-1118656, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.893866] env[63531]: DEBUG oslo_concurrency.lockutils [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "225e11e5-0558-4ad5-acd4-2a69888fe2d1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1109.894126] env[63531]: DEBUG oslo_concurrency.lockutils [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "225e11e5-0558-4ad5-acd4-2a69888fe2d1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1110.004510] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Releasing lock "refresh_cache-7fc38de1-1fbf-4312-98f6-6801d5fcd49a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1110.004845] env[63531]: DEBUG nova.compute.manager [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] Instance network_info: |[{"id": "e22c7894-d0a9-4984-993f-0c220a12dfd5", "address": "fa:16:3e:84:57:e0", "network": {"id": "80641cc0-680b-40ff-8d1b-f5a5f736a88f", "bridge": "br-int", "label": "tempest-ServersTestJSON-666288463-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "63900312a30b4942854f552758c039cf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "26472e27-9835-4f87-ab7f-ca24dfee4e83", "external-id": "nsx-vlan-transportzone-335", "segmentation_id": 335, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape22c7894-d0", "ovs_interfaceid": "e22c7894-d0a9-4984-993f-0c220a12dfd5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1110.005344] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:84:57:e0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '26472e27-9835-4f87-ab7f-ca24dfee4e83', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e22c7894-d0a9-4984-993f-0c220a12dfd5', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1110.013789] env[63531]: DEBUG oslo.service.loopingcall [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1110.014148] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1110.014436] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-277e4299-434f-41ea-9348-f2d44558a491 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.038397] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1110.038397] env[63531]: value = "task-1118657" [ 1110.038397] env[63531]: _type = "Task" [ 1110.038397] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.047436] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118657, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.103761] env[63531]: DEBUG nova.compute.manager [req-4750fbb7-7c00-4ace-a528-e08ab569ccc4 req-771b890b-f7ce-4dcc-8846-236b7a778e5a service nova] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] Received event network-vif-plugged-e22c7894-d0a9-4984-993f-0c220a12dfd5 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1110.104067] env[63531]: DEBUG oslo_concurrency.lockutils [req-4750fbb7-7c00-4ace-a528-e08ab569ccc4 req-771b890b-f7ce-4dcc-8846-236b7a778e5a service nova] Acquiring lock "7fc38de1-1fbf-4312-98f6-6801d5fcd49a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1110.104420] env[63531]: DEBUG oslo_concurrency.lockutils [req-4750fbb7-7c00-4ace-a528-e08ab569ccc4 req-771b890b-f7ce-4dcc-8846-236b7a778e5a service nova] Lock "7fc38de1-1fbf-4312-98f6-6801d5fcd49a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1110.104661] env[63531]: DEBUG oslo_concurrency.lockutils [req-4750fbb7-7c00-4ace-a528-e08ab569ccc4 req-771b890b-f7ce-4dcc-8846-236b7a778e5a service nova] Lock "7fc38de1-1fbf-4312-98f6-6801d5fcd49a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1110.104877] env[63531]: DEBUG nova.compute.manager [req-4750fbb7-7c00-4ace-a528-e08ab569ccc4 req-771b890b-f7ce-4dcc-8846-236b7a778e5a service nova] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] No waiting events found dispatching network-vif-plugged-e22c7894-d0a9-4984-993f-0c220a12dfd5 {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1110.105191] env[63531]: WARNING nova.compute.manager [req-4750fbb7-7c00-4ace-a528-e08ab569ccc4 req-771b890b-f7ce-4dcc-8846-236b7a778e5a service nova] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] Received unexpected event network-vif-plugged-e22c7894-d0a9-4984-993f-0c220a12dfd5 for instance with vm_state building and task_state spawning. [ 1110.105432] env[63531]: DEBUG nova.compute.manager [req-4750fbb7-7c00-4ace-a528-e08ab569ccc4 req-771b890b-f7ce-4dcc-8846-236b7a778e5a service nova] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] Received event network-changed-e22c7894-d0a9-4984-993f-0c220a12dfd5 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1110.105616] env[63531]: DEBUG nova.compute.manager [req-4750fbb7-7c00-4ace-a528-e08ab569ccc4 req-771b890b-f7ce-4dcc-8846-236b7a778e5a service nova] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] Refreshing instance network info cache due to event network-changed-e22c7894-d0a9-4984-993f-0c220a12dfd5. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1110.105840] env[63531]: DEBUG oslo_concurrency.lockutils [req-4750fbb7-7c00-4ace-a528-e08ab569ccc4 req-771b890b-f7ce-4dcc-8846-236b7a778e5a service nova] Acquiring lock "refresh_cache-7fc38de1-1fbf-4312-98f6-6801d5fcd49a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1110.105998] env[63531]: DEBUG oslo_concurrency.lockutils [req-4750fbb7-7c00-4ace-a528-e08ab569ccc4 req-771b890b-f7ce-4dcc-8846-236b7a778e5a service nova] Acquired lock "refresh_cache-7fc38de1-1fbf-4312-98f6-6801d5fcd49a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1110.106243] env[63531]: DEBUG nova.network.neutron [req-4750fbb7-7c00-4ace-a528-e08ab569ccc4 req-771b890b-f7ce-4dcc-8846-236b7a778e5a service nova] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] Refreshing network info cache for port e22c7894-d0a9-4984-993f-0c220a12dfd5 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1110.108464] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.400s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1110.109091] env[63531]: DEBUG nova.compute.manager [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1110.113050] env[63531]: DEBUG oslo_concurrency.lockutils [None req-de477bf0-a47e-4909-bd9d-2464242d52bb tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.342s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1110.113050] env[63531]: DEBUG nova.objects.instance [None req-de477bf0-a47e-4909-bd9d-2464242d52bb tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lazy-loading 'resources' on Instance uuid 18057005-1ff1-4e3d-bccc-8c89755d6db7 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1110.376576] env[63531]: DEBUG oslo_vmware.api [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Task: {'id': task-1118656, 'name': PowerOnVM_Task, 'duration_secs': 0.456177} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.376857] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] [instance: fbf63fa0-c8ec-41c7-878e-1446537b7b49] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1110.377079] env[63531]: INFO nova.compute.manager [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] [instance: fbf63fa0-c8ec-41c7-878e-1446537b7b49] Took 8.85 seconds to spawn the instance on the hypervisor. [ 1110.377270] env[63531]: DEBUG nova.compute.manager [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] [instance: fbf63fa0-c8ec-41c7-878e-1446537b7b49] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1110.378099] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9805be67-0d70-4d44-ab8a-99f58f4c080f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.397025] env[63531]: DEBUG nova.compute.manager [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1110.549063] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118657, 'name': CreateVM_Task, 'duration_secs': 0.339167} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.549179] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1110.549888] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1110.550227] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1110.550596] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1110.550983] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-95971d1e-d304-49d5-b4b3-71ecdbb91057 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.555941] env[63531]: DEBUG oslo_vmware.api [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 1110.555941] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]528387ba-56da-b907-7e7c-fecb2bb05f26" [ 1110.555941] env[63531]: _type = "Task" [ 1110.555941] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.566703] env[63531]: DEBUG oslo_vmware.api [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]528387ba-56da-b907-7e7c-fecb2bb05f26, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.620872] env[63531]: DEBUG nova.compute.utils [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1110.623405] env[63531]: DEBUG nova.compute.manager [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1110.623644] env[63531]: DEBUG nova.network.neutron [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1110.705261] env[63531]: DEBUG nova.policy [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '36cb2f0de54f417384fa1fb2210de4e4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8ba8b284854242c392aec5326e996239', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 1110.784953] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90ef38cd-748b-4183-9fd8-2f2e1c6ee7fa {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.799018] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44ed8859-c829-475e-b4ea-4b744388e1c0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.831094] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5af578a2-dd7a-4b46-bfb0-bb8085f7426f tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquiring lock "49967454-a4f9-47f6-a2a9-0cd0c29e18d8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1110.831364] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5af578a2-dd7a-4b46-bfb0-bb8085f7426f tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "49967454-a4f9-47f6-a2a9-0cd0c29e18d8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1110.831628] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5af578a2-dd7a-4b46-bfb0-bb8085f7426f tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquiring lock "49967454-a4f9-47f6-a2a9-0cd0c29e18d8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1110.831754] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5af578a2-dd7a-4b46-bfb0-bb8085f7426f tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "49967454-a4f9-47f6-a2a9-0cd0c29e18d8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1110.831924] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5af578a2-dd7a-4b46-bfb0-bb8085f7426f tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "49967454-a4f9-47f6-a2a9-0cd0c29e18d8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1110.837254] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-275b89c2-d206-4dd1-b22b-307b5e28f378 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.840111] env[63531]: INFO nova.compute.manager [None req-5af578a2-dd7a-4b46-bfb0-bb8085f7426f tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Terminating instance [ 1110.842735] env[63531]: DEBUG nova.compute.manager [None req-5af578a2-dd7a-4b46-bfb0-bb8085f7426f tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1110.842993] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-5af578a2-dd7a-4b46-bfb0-bb8085f7426f tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1110.843591] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ea6106db-cdfa-4251-b3ad-f8d9c518ea69 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.848772] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd901a55-fd0b-4ec6-86bf-6107f066c51f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.853825] env[63531]: DEBUG oslo_vmware.api [None req-5af578a2-dd7a-4b46-bfb0-bb8085f7426f tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 1110.853825] env[63531]: value = "task-1118658" [ 1110.853825] env[63531]: _type = "Task" [ 1110.853825] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.865292] env[63531]: DEBUG nova.compute.provider_tree [None req-de477bf0-a47e-4909-bd9d-2464242d52bb tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1110.872075] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-5af578a2-dd7a-4b46-bfb0-bb8085f7426f tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] VM already powered off {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1110.872294] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-5af578a2-dd7a-4b46-bfb0-bb8085f7426f tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Volume detach. Driver type: vmdk {{(pid=63531) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1110.872495] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-5af578a2-dd7a-4b46-bfb0-bb8085f7426f tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244800', 'volume_id': '422ea428-1dc4-40f2-8151-a5e18a596f9b', 'name': 'volume-422ea428-1dc4-40f2-8151-a5e18a596f9b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '49967454-a4f9-47f6-a2a9-0cd0c29e18d8', 'attached_at': '', 'detached_at': '', 'volume_id': '422ea428-1dc4-40f2-8151-a5e18a596f9b', 'serial': '422ea428-1dc4-40f2-8151-a5e18a596f9b'} {{(pid=63531) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1110.873719] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4327bc7b-3418-4271-96d7-0982b6e7f568 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.895646] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59826529-674c-47ee-b0e1-09913a8dea39 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.905861] env[63531]: INFO nova.compute.manager [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] [instance: fbf63fa0-c8ec-41c7-878e-1446537b7b49] Took 19.79 seconds to build instance. [ 1110.911892] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61c23926-ad23-4f43-bed6-c4e5d19a3041 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.918570] env[63531]: DEBUG oslo_concurrency.lockutils [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1110.945534] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2c5c30a-c987-490e-8c2d-d285087fcaee {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.960899] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-5af578a2-dd7a-4b46-bfb0-bb8085f7426f tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] The volume has not been displaced from its original location: [datastore2] volume-422ea428-1dc4-40f2-8151-a5e18a596f9b/volume-422ea428-1dc4-40f2-8151-a5e18a596f9b.vmdk. No consolidation needed. {{(pid=63531) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1110.966106] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-5af578a2-dd7a-4b46-bfb0-bb8085f7426f tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Reconfiguring VM instance instance-0000006b to detach disk 2000 {{(pid=63531) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1110.966943] env[63531]: DEBUG nova.network.neutron [req-4750fbb7-7c00-4ace-a528-e08ab569ccc4 req-771b890b-f7ce-4dcc-8846-236b7a778e5a service nova] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] Updated VIF entry in instance network info cache for port e22c7894-d0a9-4984-993f-0c220a12dfd5. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1110.967275] env[63531]: DEBUG nova.network.neutron [req-4750fbb7-7c00-4ace-a528-e08ab569ccc4 req-771b890b-f7ce-4dcc-8846-236b7a778e5a service nova] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] Updating instance_info_cache with network_info: [{"id": "e22c7894-d0a9-4984-993f-0c220a12dfd5", "address": "fa:16:3e:84:57:e0", "network": {"id": "80641cc0-680b-40ff-8d1b-f5a5f736a88f", "bridge": "br-int", "label": "tempest-ServersTestJSON-666288463-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "63900312a30b4942854f552758c039cf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "26472e27-9835-4f87-ab7f-ca24dfee4e83", "external-id": "nsx-vlan-transportzone-335", "segmentation_id": 335, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape22c7894-d0", "ovs_interfaceid": "e22c7894-d0a9-4984-993f-0c220a12dfd5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1110.968744] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6e341d35-8352-4abc-bf33-7595329a13a7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.982387] env[63531]: DEBUG oslo_concurrency.lockutils [req-4750fbb7-7c00-4ace-a528-e08ab569ccc4 req-771b890b-f7ce-4dcc-8846-236b7a778e5a service nova] Releasing lock "refresh_cache-7fc38de1-1fbf-4312-98f6-6801d5fcd49a" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1110.988781] env[63531]: DEBUG oslo_vmware.api [None req-5af578a2-dd7a-4b46-bfb0-bb8085f7426f tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 1110.988781] env[63531]: value = "task-1118659" [ 1110.988781] env[63531]: _type = "Task" [ 1110.988781] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.998145] env[63531]: DEBUG oslo_vmware.api [None req-5af578a2-dd7a-4b46-bfb0-bb8085f7426f tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118659, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.071572] env[63531]: DEBUG oslo_vmware.api [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]528387ba-56da-b907-7e7c-fecb2bb05f26, 'name': SearchDatastore_Task, 'duration_secs': 0.010168} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.071907] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1111.072180] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1111.074020] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1111.074020] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1111.074020] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1111.074020] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fc033df2-67cf-45b5-b4b2-122c1382ff5e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.084215] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1111.084215] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1111.084866] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d9639fe7-16e0-49c2-be01-0d65f0fa4cbc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.089253] env[63531]: DEBUG nova.network.neutron [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Successfully created port: 1d65a3d9-a09b-4266-8a05-6abd7ddd18c7 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1111.092420] env[63531]: DEBUG oslo_vmware.api [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 1111.092420] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]520be931-b891-0196-65a8-0250aef43f44" [ 1111.092420] env[63531]: _type = "Task" [ 1111.092420] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.101153] env[63531]: DEBUG oslo_vmware.api [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]520be931-b891-0196-65a8-0250aef43f44, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.127013] env[63531]: DEBUG nova.compute.manager [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1111.279328] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1111.298009] env[63531]: DEBUG oslo_concurrency.lockutils [None req-39bd932e-2b6d-481f-a2c5-d30532c36f5b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Acquiring lock "fbf63fa0-c8ec-41c7-878e-1446537b7b49" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1111.368225] env[63531]: DEBUG nova.scheduler.client.report [None req-de477bf0-a47e-4909-bd9d-2464242d52bb tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1111.407945] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3140fc9c-1064-4470-a5a2-1cde4835dc6b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Lock "fbf63fa0-c8ec-41c7-878e-1446537b7b49" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.299s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1111.408390] env[63531]: DEBUG oslo_concurrency.lockutils [None req-39bd932e-2b6d-481f-a2c5-d30532c36f5b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Lock "fbf63fa0-c8ec-41c7-878e-1446537b7b49" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.111s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1111.408735] env[63531]: DEBUG oslo_concurrency.lockutils [None req-39bd932e-2b6d-481f-a2c5-d30532c36f5b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Acquiring lock "fbf63fa0-c8ec-41c7-878e-1446537b7b49-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1111.408988] env[63531]: DEBUG oslo_concurrency.lockutils [None req-39bd932e-2b6d-481f-a2c5-d30532c36f5b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Lock "fbf63fa0-c8ec-41c7-878e-1446537b7b49-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1111.409288] env[63531]: DEBUG oslo_concurrency.lockutils [None req-39bd932e-2b6d-481f-a2c5-d30532c36f5b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Lock "fbf63fa0-c8ec-41c7-878e-1446537b7b49-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1111.411300] env[63531]: INFO nova.compute.manager [None req-39bd932e-2b6d-481f-a2c5-d30532c36f5b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] [instance: fbf63fa0-c8ec-41c7-878e-1446537b7b49] Terminating instance [ 1111.413124] env[63531]: DEBUG nova.compute.manager [None req-39bd932e-2b6d-481f-a2c5-d30532c36f5b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] [instance: fbf63fa0-c8ec-41c7-878e-1446537b7b49] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1111.413429] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-39bd932e-2b6d-481f-a2c5-d30532c36f5b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] [instance: fbf63fa0-c8ec-41c7-878e-1446537b7b49] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1111.414607] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30b6d8e3-377e-45f7-a634-59bf10bf309d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.425262] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-39bd932e-2b6d-481f-a2c5-d30532c36f5b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] [instance: fbf63fa0-c8ec-41c7-878e-1446537b7b49] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1111.425520] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f9d98889-94a4-4751-ac5e-ebcdb78c5c65 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.432210] env[63531]: DEBUG oslo_vmware.api [None req-39bd932e-2b6d-481f-a2c5-d30532c36f5b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Waiting for the task: (returnval){ [ 1111.432210] env[63531]: value = "task-1118660" [ 1111.432210] env[63531]: _type = "Task" [ 1111.432210] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.445345] env[63531]: DEBUG oslo_vmware.api [None req-39bd932e-2b6d-481f-a2c5-d30532c36f5b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Task: {'id': task-1118660, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.499410] env[63531]: DEBUG oslo_vmware.api [None req-5af578a2-dd7a-4b46-bfb0-bb8085f7426f tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118659, 'name': ReconfigVM_Task, 'duration_secs': 0.181264} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.499794] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-5af578a2-dd7a-4b46-bfb0-bb8085f7426f tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Reconfigured VM instance instance-0000006b to detach disk 2000 {{(pid=63531) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1111.504739] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4eeb593e-cabf-4911-b9c5-9a8102a32570 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.520897] env[63531]: DEBUG oslo_vmware.api [None req-5af578a2-dd7a-4b46-bfb0-bb8085f7426f tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 1111.520897] env[63531]: value = "task-1118661" [ 1111.520897] env[63531]: _type = "Task" [ 1111.520897] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.529158] env[63531]: DEBUG oslo_vmware.api [None req-5af578a2-dd7a-4b46-bfb0-bb8085f7426f tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118661, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.606050] env[63531]: DEBUG oslo_vmware.api [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]520be931-b891-0196-65a8-0250aef43f44, 'name': SearchDatastore_Task, 'duration_secs': 0.011268} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.607084] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7deffc88-0fc6-45b2-911a-1043d56393b7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.614242] env[63531]: DEBUG oslo_vmware.api [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 1111.614242] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5264a506-a317-0d05-992b-3a72ade80edc" [ 1111.614242] env[63531]: _type = "Task" [ 1111.614242] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.623736] env[63531]: DEBUG oslo_vmware.api [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5264a506-a317-0d05-992b-3a72ade80edc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.873785] env[63531]: DEBUG oslo_concurrency.lockutils [None req-de477bf0-a47e-4909-bd9d-2464242d52bb tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.761s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1111.876152] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 4.274s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1111.876345] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1111.876715] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63531) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1111.876807] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ca929e81-526c-4b49-8799-8ac6635bcbc4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.369s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1111.876991] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ca929e81-526c-4b49-8799-8ac6635bcbc4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1111.878698] env[63531]: DEBUG oslo_concurrency.lockutils [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.960s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1111.880197] env[63531]: INFO nova.compute.claims [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1111.883795] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bdd5b1b-254a-45bc-ae2e-2943c3b523cc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.892840] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0f97179-709f-42fb-a2f6-ff0089518955 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.897674] env[63531]: INFO nova.scheduler.client.report [None req-de477bf0-a47e-4909-bd9d-2464242d52bb tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Deleted allocations for instance 18057005-1ff1-4e3d-bccc-8c89755d6db7 [ 1111.899339] env[63531]: INFO nova.scheduler.client.report [None req-ca929e81-526c-4b49-8799-8ac6635bcbc4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Deleted allocations for instance 96c2910b-1e25-4053-9ad9-c155402ef0ac [ 1111.913698] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae8cf225-7cbd-48b5-b66f-1c2127592fd9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.941160] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8fce445-98ae-49d3-bb04-e8509eda178d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.948161] env[63531]: DEBUG oslo_vmware.api [None req-39bd932e-2b6d-481f-a2c5-d30532c36f5b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Task: {'id': task-1118660, 'name': PowerOffVM_Task, 'duration_secs': 0.179564} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.948694] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-39bd932e-2b6d-481f-a2c5-d30532c36f5b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] [instance: fbf63fa0-c8ec-41c7-878e-1446537b7b49] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1111.948872] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-39bd932e-2b6d-481f-a2c5-d30532c36f5b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] [instance: fbf63fa0-c8ec-41c7-878e-1446537b7b49] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1111.949157] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5a974f68-2d81-482f-a796-4d7977398851 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.978824] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180711MB free_disk=170GB free_vcpus=48 pci_devices=None {{(pid=63531) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1111.978993] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1112.032184] env[63531]: DEBUG oslo_vmware.api [None req-5af578a2-dd7a-4b46-bfb0-bb8085f7426f tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118661, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.061791] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-39bd932e-2b6d-481f-a2c5-d30532c36f5b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] [instance: fbf63fa0-c8ec-41c7-878e-1446537b7b49] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1112.062052] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-39bd932e-2b6d-481f-a2c5-d30532c36f5b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] [instance: fbf63fa0-c8ec-41c7-878e-1446537b7b49] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1112.062256] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-39bd932e-2b6d-481f-a2c5-d30532c36f5b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Deleting the datastore file [datastore1] fbf63fa0-c8ec-41c7-878e-1446537b7b49 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1112.062527] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-40eff2ff-2d8e-4688-a901-d3ab32af3283 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.069118] env[63531]: DEBUG oslo_vmware.api [None req-39bd932e-2b6d-481f-a2c5-d30532c36f5b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Waiting for the task: (returnval){ [ 1112.069118] env[63531]: value = "task-1118663" [ 1112.069118] env[63531]: _type = "Task" [ 1112.069118] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.077943] env[63531]: DEBUG oslo_vmware.api [None req-39bd932e-2b6d-481f-a2c5-d30532c36f5b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Task: {'id': task-1118663, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.124353] env[63531]: DEBUG oslo_vmware.api [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5264a506-a317-0d05-992b-3a72ade80edc, 'name': SearchDatastore_Task, 'duration_secs': 0.011372} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.124679] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1112.124987] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 7fc38de1-1fbf-4312-98f6-6801d5fcd49a/7fc38de1-1fbf-4312-98f6-6801d5fcd49a.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1112.125308] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f80047f9-c292-49a2-b32c-708e5998082c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.132518] env[63531]: DEBUG oslo_vmware.api [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 1112.132518] env[63531]: value = "task-1118664" [ 1112.132518] env[63531]: _type = "Task" [ 1112.132518] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.136643] env[63531]: DEBUG nova.compute.manager [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1112.144275] env[63531]: DEBUG oslo_vmware.api [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118664, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.144706] env[63531]: DEBUG oslo_concurrency.lockutils [None req-99659a4d-3f40-428c-9238-204dfed9bdfb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquiring lock "49967454-a4f9-47f6-a2a9-0cd0c29e18d8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1112.162078] env[63531]: DEBUG nova.virt.hardware [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1112.162413] env[63531]: DEBUG nova.virt.hardware [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1112.162624] env[63531]: DEBUG nova.virt.hardware [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1112.162861] env[63531]: DEBUG nova.virt.hardware [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1112.163066] env[63531]: DEBUG nova.virt.hardware [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1112.163233] env[63531]: DEBUG nova.virt.hardware [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1112.163452] env[63531]: DEBUG nova.virt.hardware [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1112.163640] env[63531]: DEBUG nova.virt.hardware [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1112.163810] env[63531]: DEBUG nova.virt.hardware [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1112.163981] env[63531]: DEBUG nova.virt.hardware [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1112.164182] env[63531]: DEBUG nova.virt.hardware [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1112.165062] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca7c023c-e0c3-4c69-8c1b-d004aebc2781 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.174739] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77651f4a-1f57-4fb2-8d5c-e0135c66ba09 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.409256] env[63531]: DEBUG oslo_concurrency.lockutils [None req-de477bf0-a47e-4909-bd9d-2464242d52bb tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "18057005-1ff1-4e3d-bccc-8c89755d6db7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.640s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1112.417089] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ca929e81-526c-4b49-8799-8ac6635bcbc4 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "96c2910b-1e25-4053-9ad9-c155402ef0ac" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.003s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1112.523278] env[63531]: DEBUG nova.compute.manager [req-51d57b70-d336-4cea-a0e2-f09753bbfc15 req-b97d7e3a-9713-40c2-b4f6-b01fc32ec328 service nova] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Received event network-vif-plugged-1d65a3d9-a09b-4266-8a05-6abd7ddd18c7 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1112.523541] env[63531]: DEBUG oslo_concurrency.lockutils [req-51d57b70-d336-4cea-a0e2-f09753bbfc15 req-b97d7e3a-9713-40c2-b4f6-b01fc32ec328 service nova] Acquiring lock "98853767-791e-4568-8bb8-30dfcc473ee3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1112.523784] env[63531]: DEBUG oslo_concurrency.lockutils [req-51d57b70-d336-4cea-a0e2-f09753bbfc15 req-b97d7e3a-9713-40c2-b4f6-b01fc32ec328 service nova] Lock "98853767-791e-4568-8bb8-30dfcc473ee3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1112.523967] env[63531]: DEBUG oslo_concurrency.lockutils [req-51d57b70-d336-4cea-a0e2-f09753bbfc15 req-b97d7e3a-9713-40c2-b4f6-b01fc32ec328 service nova] Lock "98853767-791e-4568-8bb8-30dfcc473ee3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1112.524170] env[63531]: DEBUG nova.compute.manager [req-51d57b70-d336-4cea-a0e2-f09753bbfc15 req-b97d7e3a-9713-40c2-b4f6-b01fc32ec328 service nova] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] No waiting events found dispatching network-vif-plugged-1d65a3d9-a09b-4266-8a05-6abd7ddd18c7 {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1112.524423] env[63531]: WARNING nova.compute.manager [req-51d57b70-d336-4cea-a0e2-f09753bbfc15 req-b97d7e3a-9713-40c2-b4f6-b01fc32ec328 service nova] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Received unexpected event network-vif-plugged-1d65a3d9-a09b-4266-8a05-6abd7ddd18c7 for instance with vm_state building and task_state spawning. [ 1112.535836] env[63531]: DEBUG oslo_vmware.api [None req-5af578a2-dd7a-4b46-bfb0-bb8085f7426f tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118661, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.582636] env[63531]: DEBUG oslo_vmware.api [None req-39bd932e-2b6d-481f-a2c5-d30532c36f5b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Task: {'id': task-1118663, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.210884} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.582981] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-39bd932e-2b6d-481f-a2c5-d30532c36f5b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1112.583256] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-39bd932e-2b6d-481f-a2c5-d30532c36f5b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] [instance: fbf63fa0-c8ec-41c7-878e-1446537b7b49] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1112.583522] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-39bd932e-2b6d-481f-a2c5-d30532c36f5b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] [instance: fbf63fa0-c8ec-41c7-878e-1446537b7b49] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1112.583729] env[63531]: INFO nova.compute.manager [None req-39bd932e-2b6d-481f-a2c5-d30532c36f5b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] [instance: fbf63fa0-c8ec-41c7-878e-1446537b7b49] Took 1.17 seconds to destroy the instance on the hypervisor. [ 1112.584044] env[63531]: DEBUG oslo.service.loopingcall [None req-39bd932e-2b6d-481f-a2c5-d30532c36f5b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1112.584292] env[63531]: DEBUG nova.compute.manager [-] [instance: fbf63fa0-c8ec-41c7-878e-1446537b7b49] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1112.584436] env[63531]: DEBUG nova.network.neutron [-] [instance: fbf63fa0-c8ec-41c7-878e-1446537b7b49] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1112.643893] env[63531]: DEBUG oslo_vmware.api [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118664, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.681587] env[63531]: DEBUG nova.network.neutron [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Successfully updated port: 1d65a3d9-a09b-4266-8a05-6abd7ddd18c7 {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1113.040260] env[63531]: DEBUG oslo_vmware.api [None req-5af578a2-dd7a-4b46-bfb0-bb8085f7426f tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118661, 'name': ReconfigVM_Task, 'duration_secs': 1.13309} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.040583] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-5af578a2-dd7a-4b46-bfb0-bb8085f7426f tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244800', 'volume_id': '422ea428-1dc4-40f2-8151-a5e18a596f9b', 'name': 'volume-422ea428-1dc4-40f2-8151-a5e18a596f9b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '49967454-a4f9-47f6-a2a9-0cd0c29e18d8', 'attached_at': '', 'detached_at': '', 'volume_id': '422ea428-1dc4-40f2-8151-a5e18a596f9b', 'serial': '422ea428-1dc4-40f2-8151-a5e18a596f9b'} {{(pid=63531) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1113.040858] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5af578a2-dd7a-4b46-bfb0-bb8085f7426f tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1113.041670] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd0daf80-3f92-44e0-9bab-b85d216120b5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.045881] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-363c36bf-ffa9-49b8-842a-1a6fb954501a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.053877] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5af578a2-dd7a-4b46-bfb0-bb8085f7426f tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1113.055684] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f4cd8ba3-c0e4-4aec-98dc-1acbff66dd03 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.058034] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a5a44c6-54d6-464d-a10b-dcfb482a61f5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.093861] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f536ccc8-66d9-4c7a-9bd0-99f2885fadb8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.101817] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cac0365c-1dc3-40d4-b1cc-bf319f8cc1e0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.116800] env[63531]: DEBUG nova.compute.provider_tree [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Updating inventory in ProviderTree for provider 9996a314-a62e-4282-8252-f387e7e3fabb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1113.139382] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5af578a2-dd7a-4b46-bfb0-bb8085f7426f tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1113.139382] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5af578a2-dd7a-4b46-bfb0-bb8085f7426f tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Deleting contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1113.139382] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-5af578a2-dd7a-4b46-bfb0-bb8085f7426f tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Deleting the datastore file [datastore2] 49967454-a4f9-47f6-a2a9-0cd0c29e18d8 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1113.139662] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-354e3ff2-c511-47e4-9dec-3f09e8e136ee {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.145617] env[63531]: DEBUG oslo_vmware.api [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118664, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.540401} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.146475] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 7fc38de1-1fbf-4312-98f6-6801d5fcd49a/7fc38de1-1fbf-4312-98f6-6801d5fcd49a.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1113.146583] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1113.146846] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-19667fe1-773a-44b2-b2db-d6d805022e42 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.151406] env[63531]: DEBUG oslo_vmware.api [None req-5af578a2-dd7a-4b46-bfb0-bb8085f7426f tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 1113.151406] env[63531]: value = "task-1118666" [ 1113.151406] env[63531]: _type = "Task" [ 1113.151406] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.156918] env[63531]: DEBUG oslo_vmware.api [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 1113.156918] env[63531]: value = "task-1118667" [ 1113.156918] env[63531]: _type = "Task" [ 1113.156918] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.163727] env[63531]: DEBUG oslo_vmware.api [None req-5af578a2-dd7a-4b46-bfb0-bb8085f7426f tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118666, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.169867] env[63531]: DEBUG oslo_vmware.api [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118667, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.184854] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "refresh_cache-98853767-791e-4568-8bb8-30dfcc473ee3" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1113.184854] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquired lock "refresh_cache-98853767-791e-4568-8bb8-30dfcc473ee3" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1113.185050] env[63531]: DEBUG nova.network.neutron [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1113.323548] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7d07d70a-ef6c-4abd-8c0f-95342d2a4723 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquiring lock "89f19669-53a9-4425-a3c6-d5f8873b707e" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1113.323841] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7d07d70a-ef6c-4abd-8c0f-95342d2a4723 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "89f19669-53a9-4425-a3c6-d5f8873b707e" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.389421] env[63531]: DEBUG nova.network.neutron [-] [instance: fbf63fa0-c8ec-41c7-878e-1446537b7b49] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1113.549175] env[63531]: DEBUG oslo_concurrency.lockutils [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquiring lock "3794f44c-8f6d-4bab-b9e0-9f6c819e1a97" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1113.549380] env[63531]: DEBUG oslo_concurrency.lockutils [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "3794f44c-8f6d-4bab-b9e0-9f6c819e1a97" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.637255] env[63531]: DEBUG oslo_concurrency.lockutils [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "2d51997d-f88c-468c-9a56-0331a06ca56c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1113.637492] env[63531]: DEBUG oslo_concurrency.lockutils [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "2d51997d-f88c-468c-9a56-0331a06ca56c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.640121] env[63531]: ERROR nova.scheduler.client.report [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [req-92232dc5-e614-419d-8e12-21c7bb083b99] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 9996a314-a62e-4282-8252-f387e7e3fabb. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-92232dc5-e614-419d-8e12-21c7bb083b99"}]} [ 1113.656676] env[63531]: DEBUG nova.scheduler.client.report [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Refreshing inventories for resource provider 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1113.665230] env[63531]: DEBUG oslo_vmware.api [None req-5af578a2-dd7a-4b46-bfb0-bb8085f7426f tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118666, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.090982} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.665900] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-5af578a2-dd7a-4b46-bfb0-bb8085f7426f tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1113.666110] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5af578a2-dd7a-4b46-bfb0-bb8085f7426f tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Deleted contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1113.666238] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5af578a2-dd7a-4b46-bfb0-bb8085f7426f tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1113.667026] env[63531]: INFO nova.compute.manager [None req-5af578a2-dd7a-4b46-bfb0-bb8085f7426f tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Took 2.82 seconds to destroy the instance on the hypervisor. [ 1113.667026] env[63531]: DEBUG oslo.service.loopingcall [None req-5af578a2-dd7a-4b46-bfb0-bb8085f7426f tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1113.667226] env[63531]: DEBUG nova.compute.manager [-] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1113.667547] env[63531]: DEBUG nova.network.neutron [-] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1113.672394] env[63531]: DEBUG oslo_vmware.api [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118667, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072378} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.673208] env[63531]: DEBUG nova.scheduler.client.report [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Updating ProviderTree inventory for provider 9996a314-a62e-4282-8252-f387e7e3fabb from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1113.673431] env[63531]: DEBUG nova.compute.provider_tree [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Updating inventory in ProviderTree for provider 9996a314-a62e-4282-8252-f387e7e3fabb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1113.675412] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1113.676325] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cd3f51b-2a31-4aa7-bd3d-64d4506af5ce {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.678769] env[63531]: DEBUG oslo_concurrency.lockutils [None req-013f893f-db1e-4ad5-ad0b-61a62d7202d5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquiring lock "c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1113.678993] env[63531]: DEBUG oslo_concurrency.lockutils [None req-013f893f-db1e-4ad5-ad0b-61a62d7202d5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.679213] env[63531]: DEBUG oslo_concurrency.lockutils [None req-013f893f-db1e-4ad5-ad0b-61a62d7202d5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquiring lock "c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1113.679402] env[63531]: DEBUG oslo_concurrency.lockutils [None req-013f893f-db1e-4ad5-ad0b-61a62d7202d5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.679570] env[63531]: DEBUG oslo_concurrency.lockutils [None req-013f893f-db1e-4ad5-ad0b-61a62d7202d5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1113.681611] env[63531]: INFO nova.compute.manager [None req-013f893f-db1e-4ad5-ad0b-61a62d7202d5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] Terminating instance [ 1113.683690] env[63531]: DEBUG nova.compute.manager [None req-013f893f-db1e-4ad5-ad0b-61a62d7202d5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1113.683884] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-013f893f-db1e-4ad5-ad0b-61a62d7202d5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1113.684703] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d8bb3da-acce-49ba-bbce-2ec61bb233a7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.688500] env[63531]: DEBUG nova.scheduler.client.report [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Refreshing aggregate associations for resource provider 9996a314-a62e-4282-8252-f387e7e3fabb, aggregates: None {{(pid=63531) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1113.709796] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] Reconfiguring VM instance instance-0000006f to attach disk [datastore1] 7fc38de1-1fbf-4312-98f6-6801d5fcd49a/7fc38de1-1fbf-4312-98f6-6801d5fcd49a.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1113.711050] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2072817c-7ea0-4f5b-87d8-bc3712d7a2b8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.727313] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-013f893f-db1e-4ad5-ad0b-61a62d7202d5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1113.728225] env[63531]: DEBUG nova.scheduler.client.report [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Refreshing trait associations for resource provider 9996a314-a62e-4282-8252-f387e7e3fabb, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE {{(pid=63531) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1113.730486] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2739a362-e8ce-4764-990f-5d0b040a02ed {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.736092] env[63531]: DEBUG oslo_vmware.api [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 1113.736092] env[63531]: value = "task-1118668" [ 1113.736092] env[63531]: _type = "Task" [ 1113.736092] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.744398] env[63531]: DEBUG oslo_vmware.api [None req-013f893f-db1e-4ad5-ad0b-61a62d7202d5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 1113.744398] env[63531]: value = "task-1118669" [ 1113.744398] env[63531]: _type = "Task" [ 1113.744398] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.761645] env[63531]: DEBUG oslo_vmware.api [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118668, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.761980] env[63531]: DEBUG oslo_vmware.api [None req-013f893f-db1e-4ad5-ad0b-61a62d7202d5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118669, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.828802] env[63531]: DEBUG nova.compute.utils [None req-7d07d70a-ef6c-4abd-8c0f-95342d2a4723 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1113.891708] env[63531]: INFO nova.compute.manager [-] [instance: fbf63fa0-c8ec-41c7-878e-1446537b7b49] Took 1.31 seconds to deallocate network for instance. [ 1113.898377] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c01b017-6d27-4c00-a747-552220e6ccd6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.908241] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fd52a3d-51a0-4dbe-a5b0-c9a25ba39b5f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.942538] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb82559a-677d-4e7f-9d1a-a63a701fa725 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.951840] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2e5a093-c2c6-42a0-8ea8-28a8c62a106e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.967963] env[63531]: DEBUG nova.compute.provider_tree [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1113.972817] env[63531]: DEBUG nova.network.neutron [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1114.052081] env[63531]: DEBUG nova.compute.manager [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1114.140411] env[63531]: DEBUG nova.compute.manager [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1114.211289] env[63531]: DEBUG nova.network.neutron [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Updating instance_info_cache with network_info: [{"id": "1d65a3d9-a09b-4266-8a05-6abd7ddd18c7", "address": "fa:16:3e:14:b8:48", "network": {"id": "caf86428-59a5-4e50-9135-2ab8473af64d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1673994696-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ba8b284854242c392aec5326e996239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d65a3d9-a0", "ovs_interfaceid": "1d65a3d9-a09b-4266-8a05-6abd7ddd18c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1114.251171] env[63531]: DEBUG oslo_vmware.api [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118668, 'name': ReconfigVM_Task, 'duration_secs': 0.294177} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.251537] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] Reconfigured VM instance instance-0000006f to attach disk [datastore1] 7fc38de1-1fbf-4312-98f6-6801d5fcd49a/7fc38de1-1fbf-4312-98f6-6801d5fcd49a.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1114.252464] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9f37521d-ecdb-40d8-9ebc-216f528e043c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.257313] env[63531]: DEBUG oslo_vmware.api [None req-013f893f-db1e-4ad5-ad0b-61a62d7202d5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118669, 'name': PowerOffVM_Task, 'duration_secs': 0.253843} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.257462] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-013f893f-db1e-4ad5-ad0b-61a62d7202d5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1114.257790] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-013f893f-db1e-4ad5-ad0b-61a62d7202d5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1114.257875] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a87129de-4a2e-4b21-b561-ca111fe4be9a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.260860] env[63531]: DEBUG oslo_vmware.api [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 1114.260860] env[63531]: value = "task-1118670" [ 1114.260860] env[63531]: _type = "Task" [ 1114.260860] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.271271] env[63531]: DEBUG oslo_vmware.api [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118670, 'name': Rename_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.331872] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7d07d70a-ef6c-4abd-8c0f-95342d2a4723 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "89f19669-53a9-4425-a3c6-d5f8873b707e" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1114.335980] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-013f893f-db1e-4ad5-ad0b-61a62d7202d5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1114.336621] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-013f893f-db1e-4ad5-ad0b-61a62d7202d5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1114.336621] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-013f893f-db1e-4ad5-ad0b-61a62d7202d5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Deleting the datastore file [datastore1] c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1114.337284] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ce3976ab-046b-4f56-9601-6215d1d0e990 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.344871] env[63531]: DEBUG oslo_vmware.api [None req-013f893f-db1e-4ad5-ad0b-61a62d7202d5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 1114.344871] env[63531]: value = "task-1118672" [ 1114.344871] env[63531]: _type = "Task" [ 1114.344871] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.353991] env[63531]: DEBUG oslo_vmware.api [None req-013f893f-db1e-4ad5-ad0b-61a62d7202d5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118672, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.398990] env[63531]: DEBUG oslo_concurrency.lockutils [None req-39bd932e-2b6d-481f-a2c5-d30532c36f5b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1114.471195] env[63531]: DEBUG nova.scheduler.client.report [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1114.580718] env[63531]: DEBUG oslo_concurrency.lockutils [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1114.605258] env[63531]: DEBUG nova.compute.manager [req-8c56a95a-62ec-4758-b1e7-bdda0ef2479c req-75d8d2cd-5e31-4da2-b04e-64e52b84c1d4 service nova] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Received event network-changed-1d65a3d9-a09b-4266-8a05-6abd7ddd18c7 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1114.605476] env[63531]: DEBUG nova.compute.manager [req-8c56a95a-62ec-4758-b1e7-bdda0ef2479c req-75d8d2cd-5e31-4da2-b04e-64e52b84c1d4 service nova] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Refreshing instance network info cache due to event network-changed-1d65a3d9-a09b-4266-8a05-6abd7ddd18c7. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1114.605678] env[63531]: DEBUG oslo_concurrency.lockutils [req-8c56a95a-62ec-4758-b1e7-bdda0ef2479c req-75d8d2cd-5e31-4da2-b04e-64e52b84c1d4 service nova] Acquiring lock "refresh_cache-98853767-791e-4568-8bb8-30dfcc473ee3" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1114.661772] env[63531]: DEBUG oslo_concurrency.lockutils [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1114.713395] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Releasing lock "refresh_cache-98853767-791e-4568-8bb8-30dfcc473ee3" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1114.713746] env[63531]: DEBUG nova.compute.manager [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Instance network_info: |[{"id": "1d65a3d9-a09b-4266-8a05-6abd7ddd18c7", "address": "fa:16:3e:14:b8:48", "network": {"id": "caf86428-59a5-4e50-9135-2ab8473af64d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1673994696-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ba8b284854242c392aec5326e996239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d65a3d9-a0", "ovs_interfaceid": "1d65a3d9-a09b-4266-8a05-6abd7ddd18c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1114.714095] env[63531]: DEBUG oslo_concurrency.lockutils [req-8c56a95a-62ec-4758-b1e7-bdda0ef2479c req-75d8d2cd-5e31-4da2-b04e-64e52b84c1d4 service nova] Acquired lock "refresh_cache-98853767-791e-4568-8bb8-30dfcc473ee3" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1114.714290] env[63531]: DEBUG nova.network.neutron [req-8c56a95a-62ec-4758-b1e7-bdda0ef2479c req-75d8d2cd-5e31-4da2-b04e-64e52b84c1d4 service nova] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Refreshing network info cache for port 1d65a3d9-a09b-4266-8a05-6abd7ddd18c7 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1114.715667] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:14:b8:48', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '91b0f7e5-0d1a-46e2-bf73-09656211dea2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1d65a3d9-a09b-4266-8a05-6abd7ddd18c7', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1114.723266] env[63531]: DEBUG oslo.service.loopingcall [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1114.724210] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1114.724455] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8e8f1df4-4f22-47fb-8146-b7a9b53143c3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.745059] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1114.745059] env[63531]: value = "task-1118673" [ 1114.745059] env[63531]: _type = "Task" [ 1114.745059] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.753673] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118673, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.769955] env[63531]: DEBUG oslo_vmware.api [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118670, 'name': Rename_Task, 'duration_secs': 0.164666} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.770374] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1114.770639] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-eccea465-0d69-4055-b9fa-703cc26915ba {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.779071] env[63531]: DEBUG oslo_vmware.api [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 1114.779071] env[63531]: value = "task-1118674" [ 1114.779071] env[63531]: _type = "Task" [ 1114.779071] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.786743] env[63531]: DEBUG oslo_vmware.api [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118674, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.857055] env[63531]: DEBUG oslo_vmware.api [None req-013f893f-db1e-4ad5-ad0b-61a62d7202d5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118672, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.136631} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.857055] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-013f893f-db1e-4ad5-ad0b-61a62d7202d5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1114.857055] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-013f893f-db1e-4ad5-ad0b-61a62d7202d5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1114.857055] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-013f893f-db1e-4ad5-ad0b-61a62d7202d5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1114.857055] env[63531]: INFO nova.compute.manager [None req-013f893f-db1e-4ad5-ad0b-61a62d7202d5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] Took 1.17 seconds to destroy the instance on the hypervisor. [ 1114.857344] env[63531]: DEBUG oslo.service.loopingcall [None req-013f893f-db1e-4ad5-ad0b-61a62d7202d5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1114.857375] env[63531]: DEBUG nova.compute.manager [-] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1114.857504] env[63531]: DEBUG nova.network.neutron [-] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1114.938019] env[63531]: DEBUG nova.network.neutron [-] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1114.979215] env[63531]: DEBUG oslo_concurrency.lockutils [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.098s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1114.979215] env[63531]: DEBUG nova.compute.manager [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1114.980415] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.701s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1115.166443] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8381adb0-8a2d-4450-b6c8-fbb5a5761ba5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.175209] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-826ad4af-ead2-45e4-a5c7-ff18bf7ccf32 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.209216] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4014345-d955-4b3b-b6b3-653f04fa8351 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.218467] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89a0a69e-bc9d-4d66-812f-4b246eb226c0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.235103] env[63531]: DEBUG nova.compute.provider_tree [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1115.256138] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118673, 'name': CreateVM_Task, 'duration_secs': 0.367278} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.256311] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1115.257032] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1115.257213] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1115.257538] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1115.257812] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cda851db-5131-4271-ae6d-7c74aa356259 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.263933] env[63531]: DEBUG oslo_vmware.api [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for the task: (returnval){ [ 1115.263933] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52d756bf-a427-18d2-2ec6-19c91376ed82" [ 1115.263933] env[63531]: _type = "Task" [ 1115.263933] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.274426] env[63531]: DEBUG oslo_vmware.api [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52d756bf-a427-18d2-2ec6-19c91376ed82, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.287981] env[63531]: DEBUG oslo_vmware.api [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118674, 'name': PowerOnVM_Task, 'duration_secs': 0.450711} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.288314] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1115.288531] env[63531]: INFO nova.compute.manager [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] Took 7.32 seconds to spawn the instance on the hypervisor. [ 1115.288714] env[63531]: DEBUG nova.compute.manager [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1115.289502] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9bca962-922f-4f43-9972-fea6d89ca9b4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.403877] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7d07d70a-ef6c-4abd-8c0f-95342d2a4723 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquiring lock "89f19669-53a9-4425-a3c6-d5f8873b707e" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1115.404168] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7d07d70a-ef6c-4abd-8c0f-95342d2a4723 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "89f19669-53a9-4425-a3c6-d5f8873b707e" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1115.404414] env[63531]: INFO nova.compute.manager [None req-7d07d70a-ef6c-4abd-8c0f-95342d2a4723 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Attaching volume 62b849f1-75ea-47e8-80dd-1e6609042e4a to /dev/sdb [ 1115.440155] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b1a9730-5193-48a5-808e-a3898cd3e34a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.442960] env[63531]: INFO nova.compute.manager [-] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Took 1.78 seconds to deallocate network for instance. [ 1115.445911] env[63531]: DEBUG nova.network.neutron [req-8c56a95a-62ec-4758-b1e7-bdda0ef2479c req-75d8d2cd-5e31-4da2-b04e-64e52b84c1d4 service nova] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Updated VIF entry in instance network info cache for port 1d65a3d9-a09b-4266-8a05-6abd7ddd18c7. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1115.446289] env[63531]: DEBUG nova.network.neutron [req-8c56a95a-62ec-4758-b1e7-bdda0ef2479c req-75d8d2cd-5e31-4da2-b04e-64e52b84c1d4 service nova] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Updating instance_info_cache with network_info: [{"id": "1d65a3d9-a09b-4266-8a05-6abd7ddd18c7", "address": "fa:16:3e:14:b8:48", "network": {"id": "caf86428-59a5-4e50-9135-2ab8473af64d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1673994696-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ba8b284854242c392aec5326e996239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d65a3d9-a0", "ovs_interfaceid": "1d65a3d9-a09b-4266-8a05-6abd7ddd18c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1115.449424] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6690772-b358-4e93-b43c-911882dce08b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.464368] env[63531]: DEBUG nova.virt.block_device [None req-7d07d70a-ef6c-4abd-8c0f-95342d2a4723 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Updating existing volume attachment record: 6d9e63ae-867b-464d-904e-c026dac57cb2 {{(pid=63531) _volume_attach /opt/stack/nova/nova/virt/block_device.py:679}} [ 1115.485049] env[63531]: DEBUG nova.compute.utils [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1115.486260] env[63531]: DEBUG nova.compute.manager [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1115.486430] env[63531]: DEBUG nova.network.neutron [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1115.523753] env[63531]: DEBUG nova.policy [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0372c22d1e9e4f66a0417643dab4a2ec', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '94a95667c00947ea9183a6307c569c90', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 1115.606761] env[63531]: DEBUG nova.network.neutron [-] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1115.738074] env[63531]: DEBUG nova.scheduler.client.report [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1115.775448] env[63531]: DEBUG oslo_vmware.api [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52d756bf-a427-18d2-2ec6-19c91376ed82, 'name': SearchDatastore_Task, 'duration_secs': 0.009644} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.775796] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1115.776057] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1115.776482] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1115.776482] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1115.776665] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1115.777598] env[63531]: DEBUG nova.network.neutron [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Successfully created port: 8b6b5f30-0bda-4851-9288-3c3a856e8a6c {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1115.779842] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d1f18b4c-846a-4ac2-911e-9132aaeb10af {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.790137] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1115.790357] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1115.791135] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6d3af925-baad-47ad-8449-4be4bc208934 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.797855] env[63531]: DEBUG oslo_vmware.api [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for the task: (returnval){ [ 1115.797855] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52a6206b-8ac8-7fdf-5ee7-edfe82457838" [ 1115.797855] env[63531]: _type = "Task" [ 1115.797855] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.811813] env[63531]: DEBUG oslo_vmware.api [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52a6206b-8ac8-7fdf-5ee7-edfe82457838, 'name': SearchDatastore_Task, 'duration_secs': 0.010685} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.814716] env[63531]: INFO nova.compute.manager [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] Took 13.16 seconds to build instance. [ 1115.818654] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8e18d89e-9f9b-4db5-be90-5e0760f4a51c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.821939] env[63531]: DEBUG oslo_vmware.api [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for the task: (returnval){ [ 1115.821939] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]523b7077-7e09-ffaf-dad4-68cab0543eab" [ 1115.821939] env[63531]: _type = "Task" [ 1115.821939] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.830959] env[63531]: DEBUG oslo_vmware.api [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]523b7077-7e09-ffaf-dad4-68cab0543eab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.949707] env[63531]: DEBUG oslo_concurrency.lockutils [req-8c56a95a-62ec-4758-b1e7-bdda0ef2479c req-75d8d2cd-5e31-4da2-b04e-64e52b84c1d4 service nova] Releasing lock "refresh_cache-98853767-791e-4568-8bb8-30dfcc473ee3" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1115.949975] env[63531]: DEBUG nova.compute.manager [req-8c56a95a-62ec-4758-b1e7-bdda0ef2479c req-75d8d2cd-5e31-4da2-b04e-64e52b84c1d4 service nova] [instance: fbf63fa0-c8ec-41c7-878e-1446537b7b49] Received event network-vif-deleted-4fa0186c-186b-4f32-8db3-dc07235c8803 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1115.950294] env[63531]: DEBUG nova.compute.manager [req-8c56a95a-62ec-4758-b1e7-bdda0ef2479c req-75d8d2cd-5e31-4da2-b04e-64e52b84c1d4 service nova] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Received event network-vif-deleted-99a45e7c-1ddb-4c35-ac69-b2c33f976019 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1115.950494] env[63531]: INFO nova.compute.manager [req-8c56a95a-62ec-4758-b1e7-bdda0ef2479c req-75d8d2cd-5e31-4da2-b04e-64e52b84c1d4 service nova] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Neutron deleted interface 99a45e7c-1ddb-4c35-ac69-b2c33f976019; detaching it from the instance and deleting it from the info cache [ 1115.950683] env[63531]: DEBUG nova.network.neutron [req-8c56a95a-62ec-4758-b1e7-bdda0ef2479c req-75d8d2cd-5e31-4da2-b04e-64e52b84c1d4 service nova] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1115.985402] env[63531]: INFO nova.compute.manager [None req-5af578a2-dd7a-4b46-bfb0-bb8085f7426f tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Took 0.54 seconds to detach 1 volumes for instance. [ 1115.987462] env[63531]: DEBUG nova.compute.manager [None req-5af578a2-dd7a-4b46-bfb0-bb8085f7426f tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Deleting volume: 422ea428-1dc4-40f2-8151-a5e18a596f9b {{(pid=63531) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 1115.989634] env[63531]: DEBUG nova.compute.manager [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1116.109956] env[63531]: INFO nova.compute.manager [-] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] Took 1.25 seconds to deallocate network for instance. [ 1116.244091] env[63531]: DEBUG oslo_concurrency.lockutils [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.264s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1116.244347] env[63531]: INFO nova.compute.manager [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Successfully reverted task state from resize_migrating on failure for instance. [ 1116.252482] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 4.273s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1116.254556] env[63531]: ERROR oslo_messaging.rpc.server [None req-a4828acd-fb6f-4d8c-95df-e06e276b421d tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Exception during message handling: AttributeError: 'NoneType' object has no attribute 'key' [ 1116.254556] env[63531]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 1116.254556] env[63531]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/server.py", line 172, in _process_incoming [ 1116.254556] env[63531]: ERROR oslo_messaging.rpc.server res = self.dispatcher.dispatch(message) [ 1116.254556] env[63531]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 309, in dispatch [ 1116.254556] env[63531]: ERROR oslo_messaging.rpc.server return self._do_dispatch(endpoint, method, ctxt, args) [ 1116.254556] env[63531]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 229, in _do_dispatch [ 1116.254556] env[63531]: ERROR oslo_messaging.rpc.server result = func(ctxt, **new_args) [ 1116.254556] env[63531]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 1116.254556] env[63531]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1116.254556] env[63531]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1116.254556] env[63531]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1116.254556] env[63531]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1116.254556] env[63531]: ERROR oslo_messaging.rpc.server raise self.value [ 1116.254556] env[63531]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 1116.254556] env[63531]: ERROR oslo_messaging.rpc.server return f(self, context, *args, **kw) [ 1116.254556] env[63531]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 166, in decorated_function [ 1116.254556] env[63531]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1116.254556] env[63531]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1116.254556] env[63531]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1116.254556] env[63531]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1116.254556] env[63531]: ERROR oslo_messaging.rpc.server raise self.value [ 1116.254556] env[63531]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 157, in decorated_function [ 1116.254556] env[63531]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1116.254556] env[63531]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/utils.py", line 1453, in decorated_function [ 1116.254556] env[63531]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1116.254556] env[63531]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 213, in decorated_function [ 1116.254556] env[63531]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1116.254556] env[63531]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1116.254556] env[63531]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1116.254556] env[63531]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1116.254556] env[63531]: ERROR oslo_messaging.rpc.server raise self.value [ 1116.254556] env[63531]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 203, in decorated_function [ 1116.254556] env[63531]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1116.254556] env[63531]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 6071, in resize_instance [ 1116.254556] env[63531]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1116.254556] env[63531]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1116.254556] env[63531]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1116.254556] env[63531]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1116.254556] env[63531]: ERROR oslo_messaging.rpc.server raise self.value [ 1116.254556] env[63531]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 6068, in resize_instance [ 1116.254556] env[63531]: ERROR oslo_messaging.rpc.server self._resize_instance(context, instance, image, migration, [ 1116.254556] env[63531]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 6105, in _resize_instance [ 1116.254556] env[63531]: ERROR oslo_messaging.rpc.server disk_info = self.driver.migrate_disk_and_power_off( [ 1116.254556] env[63531]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 264, in migrate_disk_and_power_off [ 1116.254556] env[63531]: ERROR oslo_messaging.rpc.server return self._vmops.migrate_disk_and_power_off(context, instance, [ 1116.254556] env[63531]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1467, in migrate_disk_and_power_off [ 1116.254556] env[63531]: ERROR oslo_messaging.rpc.server self._resize_disk(instance, vm_ref, vmdk, flavor) [ 1116.254556] env[63531]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1398, in _resize_disk [ 1116.254556] env[63531]: ERROR oslo_messaging.rpc.server self._volumeops.detach_disk_from_vm(vm_ref, instance, vmdk.device) [ 1116.254556] env[63531]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 121, in detach_disk_from_vm [ 1116.254556] env[63531]: ERROR oslo_messaging.rpc.server disk_key = device.key [ 1116.254556] env[63531]: ERROR oslo_messaging.rpc.server AttributeError: 'NoneType' object has no attribute 'key' [ 1116.256042] env[63531]: ERROR oslo_messaging.rpc.server [ 1116.319460] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f403f465-55f3-4b19-b24a-92d06fa290f3 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "7fc38de1-1fbf-4312-98f6-6801d5fcd49a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.680s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1116.332487] env[63531]: DEBUG oslo_vmware.api [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]523b7077-7e09-ffaf-dad4-68cab0543eab, 'name': SearchDatastore_Task, 'duration_secs': 0.010507} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.332762] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1116.333037] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 98853767-791e-4568-8bb8-30dfcc473ee3/98853767-791e-4568-8bb8-30dfcc473ee3.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1116.333305] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-113721f9-5d8a-476d-a8b2-ff6a87036f27 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.339761] env[63531]: DEBUG oslo_vmware.api [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for the task: (returnval){ [ 1116.339761] env[63531]: value = "task-1118677" [ 1116.339761] env[63531]: _type = "Task" [ 1116.339761] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.348199] env[63531]: DEBUG oslo_vmware.api [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118677, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.455029] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dd0a0338-0619-49d6-97ed-7b03d3c6f0b4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.465022] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df695a3d-87ea-4e27-9c8d-41b4ad0bb110 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.500688] env[63531]: DEBUG nova.compute.manager [req-8c56a95a-62ec-4758-b1e7-bdda0ef2479c req-75d8d2cd-5e31-4da2-b04e-64e52b84c1d4 service nova] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Detach interface failed, port_id=99a45e7c-1ddb-4c35-ac69-b2c33f976019, reason: Instance 49967454-a4f9-47f6-a2a9-0cd0c29e18d8 could not be found. {{(pid=63531) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1116.555308] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5af578a2-dd7a-4b46-bfb0-bb8085f7426f tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1116.616728] env[63531]: DEBUG oslo_concurrency.lockutils [None req-013f893f-db1e-4ad5-ad0b-61a62d7202d5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1116.635543] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c1cd9b07-8e01-4aa4-8dc6-edf20e3503a0 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "7fc38de1-1fbf-4312-98f6-6801d5fcd49a" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1116.635815] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c1cd9b07-8e01-4aa4-8dc6-edf20e3503a0 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "7fc38de1-1fbf-4312-98f6-6801d5fcd49a" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1116.636023] env[63531]: DEBUG nova.compute.manager [None req-c1cd9b07-8e01-4aa4-8dc6-edf20e3503a0 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1116.637224] env[63531]: DEBUG nova.compute.manager [req-685f5fab-4d36-4818-bc06-f95846d722e5 req-14872e9b-67ba-41d1-abdf-0930d502e806 service nova] [instance: c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9] Received event network-vif-deleted-b8e8e4e6-84b5-46b4-9b14-65eb203d51cb {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1116.638484] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78ae030b-8c2f-4f5a-bf3b-57487e6b286f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.647783] env[63531]: DEBUG nova.compute.manager [None req-c1cd9b07-8e01-4aa4-8dc6-edf20e3503a0 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63531) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1116.648475] env[63531]: DEBUG nova.objects.instance [None req-c1cd9b07-8e01-4aa4-8dc6-edf20e3503a0 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lazy-loading 'flavor' on Instance uuid 7fc38de1-1fbf-4312-98f6-6801d5fcd49a {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1116.849784] env[63531]: DEBUG oslo_vmware.api [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118677, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.003755] env[63531]: DEBUG nova.compute.manager [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1117.031757] env[63531]: DEBUG nova.virt.hardware [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1117.032027] env[63531]: DEBUG nova.virt.hardware [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1117.032211] env[63531]: DEBUG nova.virt.hardware [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1117.032405] env[63531]: DEBUG nova.virt.hardware [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1117.032572] env[63531]: DEBUG nova.virt.hardware [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1117.032727] env[63531]: DEBUG nova.virt.hardware [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1117.032958] env[63531]: DEBUG nova.virt.hardware [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1117.033154] env[63531]: DEBUG nova.virt.hardware [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1117.033336] env[63531]: DEBUG nova.virt.hardware [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1117.033509] env[63531]: DEBUG nova.virt.hardware [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1117.033704] env[63531]: DEBUG nova.virt.hardware [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1117.034630] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c286f9cd-aa9b-42da-aef2-cd7874020176 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.043551] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54ab8ece-6ee3-497e-ab4e-1f4960dfd4c8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.153759] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1cd9b07-8e01-4aa4-8dc6-edf20e3503a0 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1117.154048] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a870f64d-e58a-4ae8-934f-06fe7bb39c8a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.161716] env[63531]: DEBUG oslo_vmware.api [None req-c1cd9b07-8e01-4aa4-8dc6-edf20e3503a0 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 1117.161716] env[63531]: value = "task-1118678" [ 1117.161716] env[63531]: _type = "Task" [ 1117.161716] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.171020] env[63531]: DEBUG oslo_vmware.api [None req-c1cd9b07-8e01-4aa4-8dc6-edf20e3503a0 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118678, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.262137] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Applying migration context for instance 49967454-a4f9-47f6-a2a9-0cd0c29e18d8 as it has an incoming, in-progress migration 72e41fa6-b1eb-48d8-9701-d5fc7cd368c7. Migration status is error {{(pid=63531) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1117.263303] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Skipping migration as instance is neither resizing nor live-migrating. {{(pid=63531) _update_usage_from_migrations /opt/stack/nova/nova/compute/resource_tracker.py:1563}} [ 1117.276950] env[63531]: DEBUG nova.network.neutron [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Successfully updated port: 8b6b5f30-0bda-4851-9288-3c3a856e8a6c {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1117.288816] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 8dec0cac-defc-42f2-ab56-6b3ae60ad858 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1117.289036] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1117.290046] env[63531]: WARNING nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1117.290046] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 89f19669-53a9-4425-a3c6-d5f8873b707e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1117.290046] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance ccc8f366-5849-45a6-9745-ff9b2be9c3bb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1117.290046] env[63531]: WARNING nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance fbf63fa0-c8ec-41c7-878e-1446537b7b49 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1117.290046] env[63531]: WARNING nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 49967454-a4f9-47f6-a2a9-0cd0c29e18d8 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'VCPU': 1, 'MEMORY_MB': 192}}. Skipping heal of allocation because we do not know what to do. [ 1117.290272] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 7fc38de1-1fbf-4312-98f6-6801d5fcd49a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1117.290272] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 98853767-791e-4568-8bb8-30dfcc473ee3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1117.290360] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 225e11e5-0558-4ad5-acd4-2a69888fe2d1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1117.350961] env[63531]: DEBUG oslo_vmware.api [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118677, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.516609} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.351772] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 98853767-791e-4568-8bb8-30dfcc473ee3/98853767-791e-4568-8bb8-30dfcc473ee3.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1117.351772] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1117.351926] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-098cb34a-5450-4225-9cd3-e6244a8478dc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.359372] env[63531]: DEBUG oslo_vmware.api [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for the task: (returnval){ [ 1117.359372] env[63531]: value = "task-1118679" [ 1117.359372] env[63531]: _type = "Task" [ 1117.359372] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.369102] env[63531]: DEBUG oslo_vmware.api [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118679, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.671994] env[63531]: DEBUG oslo_vmware.api [None req-c1cd9b07-8e01-4aa4-8dc6-edf20e3503a0 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118678, 'name': PowerOffVM_Task, 'duration_secs': 0.206311} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.672198] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1cd9b07-8e01-4aa4-8dc6-edf20e3503a0 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1117.672469] env[63531]: DEBUG nova.compute.manager [None req-c1cd9b07-8e01-4aa4-8dc6-edf20e3503a0 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1117.673304] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25a5a53d-6ccc-4a2b-a10c-29eaa5c485a4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.779907] env[63531]: DEBUG oslo_concurrency.lockutils [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "refresh_cache-225e11e5-0558-4ad5-acd4-2a69888fe2d1" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1117.780284] env[63531]: DEBUG oslo_concurrency.lockutils [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquired lock "refresh_cache-225e11e5-0558-4ad5-acd4-2a69888fe2d1" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1117.780351] env[63531]: DEBUG nova.network.neutron [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1117.793558] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1117.870242] env[63531]: DEBUG oslo_vmware.api [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118679, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068507} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.870969] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1117.871366] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2d75a36-3cdf-49d0-aa75-abe363f53584 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.895252] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Reconfiguring VM instance instance-00000070 to attach disk [datastore1] 98853767-791e-4568-8bb8-30dfcc473ee3/98853767-791e-4568-8bb8-30dfcc473ee3.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1117.895559] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-90b8df12-2da9-4522-9549-f5d18ddf6907 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.914931] env[63531]: DEBUG oslo_vmware.api [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for the task: (returnval){ [ 1117.914931] env[63531]: value = "task-1118681" [ 1117.914931] env[63531]: _type = "Task" [ 1117.914931] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.935824] env[63531]: DEBUG oslo_vmware.api [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118681, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.186712] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c1cd9b07-8e01-4aa4-8dc6-edf20e3503a0 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "7fc38de1-1fbf-4312-98f6-6801d5fcd49a" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.551s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1118.295862] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 2d51997d-f88c-468c-9a56-0331a06ca56c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1118.296120] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Total usable vcpus: 48, total allocated vcpus: 7 {{(pid=63531) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1118.296270] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1856MB phys_disk=200GB used_disk=7GB total_vcpus=48 used_vcpus=7 pci_stats=[] {{(pid=63531) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1118.312399] env[63531]: DEBUG nova.network.neutron [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1118.430231] env[63531]: DEBUG oslo_vmware.api [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118681, 'name': ReconfigVM_Task, 'duration_secs': 0.2968} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.430543] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Reconfigured VM instance instance-00000070 to attach disk [datastore1] 98853767-791e-4568-8bb8-30dfcc473ee3/98853767-791e-4568-8bb8-30dfcc473ee3.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1118.431469] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f022ba25-3e53-43ad-b833-ef2dfacfc881 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.441745] env[63531]: DEBUG oslo_vmware.api [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for the task: (returnval){ [ 1118.441745] env[63531]: value = "task-1118682" [ 1118.441745] env[63531]: _type = "Task" [ 1118.441745] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.449693] env[63531]: DEBUG oslo_vmware.api [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118682, 'name': Rename_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.451288] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17addc82-054a-4b3e-b9a2-5c064f6c2ab5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.458297] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34de56ac-33a6-4d71-97f9-4b89caaa2919 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.491077] env[63531]: DEBUG nova.network.neutron [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Updating instance_info_cache with network_info: [{"id": "8b6b5f30-0bda-4851-9288-3c3a856e8a6c", "address": "fa:16:3e:8d:e8:d4", "network": {"id": "ea2d47d3-a502-4afd-9bc5-ee4a37017901", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1361282648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94a95667c00947ea9183a6307c569c90", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6c45fd-e930-495a-9cb7-df84eda443b1", "external-id": "nsx-vlan-transportzone-407", "segmentation_id": 407, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8b6b5f30-0b", "ovs_interfaceid": "8b6b5f30-0bda-4851-9288-3c3a856e8a6c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1118.492972] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73a2b292-4691-4efc-a5c0-1ba3f4c40760 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.501335] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da787637-4311-4447-85a2-5b8649078a30 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.515789] env[63531]: DEBUG nova.compute.provider_tree [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1118.661617] env[63531]: DEBUG nova.compute.manager [req-29689757-e77e-46f3-a9b7-38bb4917ca8b req-3e2ad200-41b2-488b-8dcb-185a6ee04a52 service nova] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Received event network-vif-plugged-8b6b5f30-0bda-4851-9288-3c3a856e8a6c {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1118.661848] env[63531]: DEBUG oslo_concurrency.lockutils [req-29689757-e77e-46f3-a9b7-38bb4917ca8b req-3e2ad200-41b2-488b-8dcb-185a6ee04a52 service nova] Acquiring lock "225e11e5-0558-4ad5-acd4-2a69888fe2d1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1118.662071] env[63531]: DEBUG oslo_concurrency.lockutils [req-29689757-e77e-46f3-a9b7-38bb4917ca8b req-3e2ad200-41b2-488b-8dcb-185a6ee04a52 service nova] Lock "225e11e5-0558-4ad5-acd4-2a69888fe2d1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1118.662259] env[63531]: DEBUG oslo_concurrency.lockutils [req-29689757-e77e-46f3-a9b7-38bb4917ca8b req-3e2ad200-41b2-488b-8dcb-185a6ee04a52 service nova] Lock "225e11e5-0558-4ad5-acd4-2a69888fe2d1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1118.662460] env[63531]: DEBUG nova.compute.manager [req-29689757-e77e-46f3-a9b7-38bb4917ca8b req-3e2ad200-41b2-488b-8dcb-185a6ee04a52 service nova] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] No waiting events found dispatching network-vif-plugged-8b6b5f30-0bda-4851-9288-3c3a856e8a6c {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1118.662635] env[63531]: WARNING nova.compute.manager [req-29689757-e77e-46f3-a9b7-38bb4917ca8b req-3e2ad200-41b2-488b-8dcb-185a6ee04a52 service nova] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Received unexpected event network-vif-plugged-8b6b5f30-0bda-4851-9288-3c3a856e8a6c for instance with vm_state building and task_state spawning. [ 1118.662821] env[63531]: DEBUG nova.compute.manager [req-29689757-e77e-46f3-a9b7-38bb4917ca8b req-3e2ad200-41b2-488b-8dcb-185a6ee04a52 service nova] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Received event network-changed-8b6b5f30-0bda-4851-9288-3c3a856e8a6c {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1118.663026] env[63531]: DEBUG nova.compute.manager [req-29689757-e77e-46f3-a9b7-38bb4917ca8b req-3e2ad200-41b2-488b-8dcb-185a6ee04a52 service nova] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Refreshing instance network info cache due to event network-changed-8b6b5f30-0bda-4851-9288-3c3a856e8a6c. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1118.663212] env[63531]: DEBUG oslo_concurrency.lockutils [req-29689757-e77e-46f3-a9b7-38bb4917ca8b req-3e2ad200-41b2-488b-8dcb-185a6ee04a52 service nova] Acquiring lock "refresh_cache-225e11e5-0558-4ad5-acd4-2a69888fe2d1" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1118.950336] env[63531]: DEBUG oslo_vmware.api [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118682, 'name': Rename_Task, 'duration_secs': 0.346802} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.950649] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1118.950843] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c671ea7a-39a4-4a6e-8fd6-2e2b38bc40d6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.957629] env[63531]: DEBUG oslo_vmware.api [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for the task: (returnval){ [ 1118.957629] env[63531]: value = "task-1118683" [ 1118.957629] env[63531]: _type = "Task" [ 1118.957629] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.966717] env[63531]: DEBUG oslo_vmware.api [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118683, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.996556] env[63531]: DEBUG oslo_concurrency.lockutils [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Releasing lock "refresh_cache-225e11e5-0558-4ad5-acd4-2a69888fe2d1" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1118.996932] env[63531]: DEBUG nova.compute.manager [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Instance network_info: |[{"id": "8b6b5f30-0bda-4851-9288-3c3a856e8a6c", "address": "fa:16:3e:8d:e8:d4", "network": {"id": "ea2d47d3-a502-4afd-9bc5-ee4a37017901", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1361282648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94a95667c00947ea9183a6307c569c90", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6c45fd-e930-495a-9cb7-df84eda443b1", "external-id": "nsx-vlan-transportzone-407", "segmentation_id": 407, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8b6b5f30-0b", "ovs_interfaceid": "8b6b5f30-0bda-4851-9288-3c3a856e8a6c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1118.997291] env[63531]: DEBUG oslo_concurrency.lockutils [req-29689757-e77e-46f3-a9b7-38bb4917ca8b req-3e2ad200-41b2-488b-8dcb-185a6ee04a52 service nova] Acquired lock "refresh_cache-225e11e5-0558-4ad5-acd4-2a69888fe2d1" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1118.997483] env[63531]: DEBUG nova.network.neutron [req-29689757-e77e-46f3-a9b7-38bb4917ca8b req-3e2ad200-41b2-488b-8dcb-185a6ee04a52 service nova] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Refreshing network info cache for port 8b6b5f30-0bda-4851-9288-3c3a856e8a6c {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1118.999138] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8d:e8:d4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2b6c45fd-e930-495a-9cb7-df84eda443b1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8b6b5f30-0bda-4851-9288-3c3a856e8a6c', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1119.007448] env[63531]: DEBUG oslo.service.loopingcall [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1119.010938] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1119.011712] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-91749991-9a06-4658-949b-5ef3b4e8864d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.027800] env[63531]: DEBUG nova.scheduler.client.report [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1119.036674] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1119.036674] env[63531]: value = "task-1118684" [ 1119.036674] env[63531]: _type = "Task" [ 1119.036674] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.045591] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118684, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.243304] env[63531]: DEBUG nova.network.neutron [req-29689757-e77e-46f3-a9b7-38bb4917ca8b req-3e2ad200-41b2-488b-8dcb-185a6ee04a52 service nova] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Updated VIF entry in instance network info cache for port 8b6b5f30-0bda-4851-9288-3c3a856e8a6c. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1119.243688] env[63531]: DEBUG nova.network.neutron [req-29689757-e77e-46f3-a9b7-38bb4917ca8b req-3e2ad200-41b2-488b-8dcb-185a6ee04a52 service nova] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Updating instance_info_cache with network_info: [{"id": "8b6b5f30-0bda-4851-9288-3c3a856e8a6c", "address": "fa:16:3e:8d:e8:d4", "network": {"id": "ea2d47d3-a502-4afd-9bc5-ee4a37017901", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1361282648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94a95667c00947ea9183a6307c569c90", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6c45fd-e930-495a-9cb7-df84eda443b1", "external-id": "nsx-vlan-transportzone-407", "segmentation_id": 407, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8b6b5f30-0b", "ovs_interfaceid": "8b6b5f30-0bda-4851-9288-3c3a856e8a6c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1119.381709] env[63531]: DEBUG oslo_concurrency.lockutils [None req-46a89a78-698b-4568-a636-317e94673b85 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "7fc38de1-1fbf-4312-98f6-6801d5fcd49a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1119.382031] env[63531]: DEBUG oslo_concurrency.lockutils [None req-46a89a78-698b-4568-a636-317e94673b85 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "7fc38de1-1fbf-4312-98f6-6801d5fcd49a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1119.382269] env[63531]: DEBUG oslo_concurrency.lockutils [None req-46a89a78-698b-4568-a636-317e94673b85 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "7fc38de1-1fbf-4312-98f6-6801d5fcd49a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1119.382481] env[63531]: DEBUG oslo_concurrency.lockutils [None req-46a89a78-698b-4568-a636-317e94673b85 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "7fc38de1-1fbf-4312-98f6-6801d5fcd49a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1119.382675] env[63531]: DEBUG oslo_concurrency.lockutils [None req-46a89a78-698b-4568-a636-317e94673b85 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "7fc38de1-1fbf-4312-98f6-6801d5fcd49a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1119.385127] env[63531]: INFO nova.compute.manager [None req-46a89a78-698b-4568-a636-317e94673b85 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] Terminating instance [ 1119.387300] env[63531]: DEBUG nova.compute.manager [None req-46a89a78-698b-4568-a636-317e94673b85 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1119.387486] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-46a89a78-698b-4568-a636-317e94673b85 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1119.388591] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b280f75-9a2c-4ddc-87f5-560398bd03ae {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.397087] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-46a89a78-698b-4568-a636-317e94673b85 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1119.397628] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8167abed-5642-42bb-b5fe-d65fc35633c5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.467950] env[63531]: DEBUG oslo_vmware.api [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118683, 'name': PowerOnVM_Task, 'duration_secs': 0.484193} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.468260] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1119.468470] env[63531]: INFO nova.compute.manager [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Took 7.33 seconds to spawn the instance on the hypervisor. [ 1119.468653] env[63531]: DEBUG nova.compute.manager [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1119.469428] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83e689aa-98e3-4068-89ab-50be502c98d7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.484723] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-46a89a78-698b-4568-a636-317e94673b85 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1119.484935] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-46a89a78-698b-4568-a636-317e94673b85 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1119.485169] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-46a89a78-698b-4568-a636-317e94673b85 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Deleting the datastore file [datastore1] 7fc38de1-1fbf-4312-98f6-6801d5fcd49a {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1119.485437] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-36ce996a-ff7b-4203-941e-984f60a60e7f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.493803] env[63531]: DEBUG oslo_vmware.api [None req-46a89a78-698b-4568-a636-317e94673b85 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 1119.493803] env[63531]: value = "task-1118686" [ 1119.493803] env[63531]: _type = "Task" [ 1119.493803] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.502524] env[63531]: DEBUG oslo_vmware.api [None req-46a89a78-698b-4568-a636-317e94673b85 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118686, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.533320] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63531) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1119.533625] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.281s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1119.533949] env[63531]: DEBUG oslo_concurrency.lockutils [None req-39bd932e-2b6d-481f-a2c5-d30532c36f5b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.135s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1119.534238] env[63531]: DEBUG oslo_concurrency.lockutils [None req-39bd932e-2b6d-481f-a2c5-d30532c36f5b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1119.536859] env[63531]: DEBUG oslo_concurrency.lockutils [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.956s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1119.538937] env[63531]: INFO nova.compute.claims [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1119.542205] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1119.542454] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Cleaning up deleted instances {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 1119.555518] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118684, 'name': CreateVM_Task, 'duration_secs': 0.339915} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.555720] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1119.556525] env[63531]: DEBUG oslo_concurrency.lockutils [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1119.556626] env[63531]: DEBUG oslo_concurrency.lockutils [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1119.556973] env[63531]: DEBUG oslo_concurrency.lockutils [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1119.557261] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1eab0f3c-6ed3-474f-8932-cb4912869343 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.563257] env[63531]: DEBUG oslo_vmware.api [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 1119.563257] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5247299e-28fb-9aba-7d9a-bc977f5e44f9" [ 1119.563257] env[63531]: _type = "Task" [ 1119.563257] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.573904] env[63531]: DEBUG oslo_vmware.api [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5247299e-28fb-9aba-7d9a-bc977f5e44f9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.575080] env[63531]: INFO nova.scheduler.client.report [None req-39bd932e-2b6d-481f-a2c5-d30532c36f5b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Deleted allocations for instance fbf63fa0-c8ec-41c7-878e-1446537b7b49 [ 1119.749539] env[63531]: DEBUG oslo_concurrency.lockutils [req-29689757-e77e-46f3-a9b7-38bb4917ca8b req-3e2ad200-41b2-488b-8dcb-185a6ee04a52 service nova] Releasing lock "refresh_cache-225e11e5-0558-4ad5-acd4-2a69888fe2d1" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1119.986307] env[63531]: INFO nova.compute.manager [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Took 14.76 seconds to build instance. [ 1120.006960] env[63531]: DEBUG oslo_vmware.api [None req-46a89a78-698b-4568-a636-317e94673b85 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118686, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.130794} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.009019] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-7d07d70a-ef6c-4abd-8c0f-95342d2a4723 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Volume attach. Driver type: vmdk {{(pid=63531) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1120.009393] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-7d07d70a-ef6c-4abd-8c0f-95342d2a4723 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244812', 'volume_id': '62b849f1-75ea-47e8-80dd-1e6609042e4a', 'name': 'volume-62b849f1-75ea-47e8-80dd-1e6609042e4a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '89f19669-53a9-4425-a3c6-d5f8873b707e', 'attached_at': '', 'detached_at': '', 'volume_id': '62b849f1-75ea-47e8-80dd-1e6609042e4a', 'serial': '62b849f1-75ea-47e8-80dd-1e6609042e4a'} {{(pid=63531) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1120.009704] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-46a89a78-698b-4568-a636-317e94673b85 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1120.009935] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-46a89a78-698b-4568-a636-317e94673b85 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1120.010264] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-46a89a78-698b-4568-a636-317e94673b85 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1120.010476] env[63531]: INFO nova.compute.manager [None req-46a89a78-698b-4568-a636-317e94673b85 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] Took 0.62 seconds to destroy the instance on the hypervisor. [ 1120.010735] env[63531]: DEBUG oslo.service.loopingcall [None req-46a89a78-698b-4568-a636-317e94673b85 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1120.011813] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8ad9fbc-caae-4f30-9bec-2b1ab693aec9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.014531] env[63531]: DEBUG nova.compute.manager [-] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1120.014636] env[63531]: DEBUG nova.network.neutron [-] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1120.038537] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9d53026-4ecf-464d-a232-e6a21769af24 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.075307] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] There are 58 instances to clean {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 1120.075477] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 148c6ecd-354e-4076-b52b-c13a2119da7e] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1120.085026] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-7d07d70a-ef6c-4abd-8c0f-95342d2a4723 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Reconfiguring VM instance instance-00000069 to attach disk [datastore2] volume-62b849f1-75ea-47e8-80dd-1e6609042e4a/volume-62b849f1-75ea-47e8-80dd-1e6609042e4a.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1120.090798] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2a4adc67-9f12-416e-842c-b2cda5ad1cdd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.105542] env[63531]: DEBUG oslo_concurrency.lockutils [None req-39bd932e-2b6d-481f-a2c5-d30532c36f5b tempest-ServerAddressesTestJSON-204193486 tempest-ServerAddressesTestJSON-204193486-project-member] Lock "fbf63fa0-c8ec-41c7-878e-1446537b7b49" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.697s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1120.118853] env[63531]: DEBUG oslo_vmware.api [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5247299e-28fb-9aba-7d9a-bc977f5e44f9, 'name': SearchDatastore_Task, 'duration_secs': 0.010639} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.120306] env[63531]: DEBUG oslo_concurrency.lockutils [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1120.120562] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1120.120818] env[63531]: DEBUG oslo_concurrency.lockutils [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1120.120981] env[63531]: DEBUG oslo_concurrency.lockutils [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1120.121185] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1120.121786] env[63531]: DEBUG oslo_vmware.api [None req-7d07d70a-ef6c-4abd-8c0f-95342d2a4723 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for the task: (returnval){ [ 1120.121786] env[63531]: value = "task-1118687" [ 1120.121786] env[63531]: _type = "Task" [ 1120.121786] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.122105] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-469393a3-b00b-49cc-b526-864cfc843228 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.136709] env[63531]: DEBUG oslo_vmware.api [None req-7d07d70a-ef6c-4abd-8c0f-95342d2a4723 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118687, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.137994] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1120.138238] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1120.139317] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-05043198-3d24-474c-9d68-eea41b4d699a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.148217] env[63531]: DEBUG oslo_vmware.api [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 1120.148217] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]521821da-4999-2096-7d89-f48452488daa" [ 1120.148217] env[63531]: _type = "Task" [ 1120.148217] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.157477] env[63531]: DEBUG oslo_vmware.api [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]521821da-4999-2096-7d89-f48452488daa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.267227] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dc7cd4f-9851-4405-acc8-d7169c862a8f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.275624] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61e09831-577c-4ca7-87c5-66711c7221a4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.313538] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46b07ec1-6aff-4c08-b98f-b13049c4a2f1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.324610] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70ad7520-f681-42ff-ba43-f70dc9135a7d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.340414] env[63531]: DEBUG nova.compute.provider_tree [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1120.488325] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6bf00fcb-eb7a-4185-abe3-1d0d628a2103 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "98853767-791e-4568-8bb8-30dfcc473ee3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.267s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1120.590120] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: b5f51306-6a9a-44fa-8c1e-cc5b7ab65d16] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1120.634763] env[63531]: DEBUG oslo_vmware.api [None req-7d07d70a-ef6c-4abd-8c0f-95342d2a4723 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118687, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.658795] env[63531]: DEBUG oslo_vmware.api [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]521821da-4999-2096-7d89-f48452488daa, 'name': SearchDatastore_Task, 'duration_secs': 0.015806} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.659278] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f1ce328f-a2f6-45a4-afa9-5f8b055ad519 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.665453] env[63531]: DEBUG oslo_vmware.api [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 1120.665453] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5203cf83-60de-07f1-3dbf-08189c2e062a" [ 1120.665453] env[63531]: _type = "Task" [ 1120.665453] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.673886] env[63531]: DEBUG oslo_vmware.api [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5203cf83-60de-07f1-3dbf-08189c2e062a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.691990] env[63531]: DEBUG nova.compute.manager [req-6ff58118-8023-4a74-a7cc-c7bdabe51ad8 req-3727232a-2d22-4665-940d-8067d9e1b247 service nova] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] Received event network-vif-deleted-e22c7894-d0a9-4984-993f-0c220a12dfd5 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1120.692209] env[63531]: INFO nova.compute.manager [req-6ff58118-8023-4a74-a7cc-c7bdabe51ad8 req-3727232a-2d22-4665-940d-8067d9e1b247 service nova] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] Neutron deleted interface e22c7894-d0a9-4984-993f-0c220a12dfd5; detaching it from the instance and deleting it from the info cache [ 1120.692389] env[63531]: DEBUG nova.network.neutron [req-6ff58118-8023-4a74-a7cc-c7bdabe51ad8 req-3727232a-2d22-4665-940d-8067d9e1b247 service nova] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1120.820206] env[63531]: DEBUG nova.network.neutron [-] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1120.843603] env[63531]: DEBUG nova.scheduler.client.report [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1121.097204] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 85058146-ba4c-4e9a-bbe7-048ed56167ab] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1121.139248] env[63531]: DEBUG oslo_vmware.api [None req-7d07d70a-ef6c-4abd-8c0f-95342d2a4723 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118687, 'name': ReconfigVM_Task, 'duration_secs': 0.52983} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.139248] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-7d07d70a-ef6c-4abd-8c0f-95342d2a4723 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Reconfigured VM instance instance-00000069 to attach disk [datastore2] volume-62b849f1-75ea-47e8-80dd-1e6609042e4a/volume-62b849f1-75ea-47e8-80dd-1e6609042e4a.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1121.142960] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e4529ef8-b637-4bf3-97c1-7d613a88b455 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.158853] env[63531]: DEBUG oslo_vmware.api [None req-7d07d70a-ef6c-4abd-8c0f-95342d2a4723 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for the task: (returnval){ [ 1121.158853] env[63531]: value = "task-1118688" [ 1121.158853] env[63531]: _type = "Task" [ 1121.158853] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.170917] env[63531]: DEBUG oslo_vmware.api [None req-7d07d70a-ef6c-4abd-8c0f-95342d2a4723 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118688, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.178997] env[63531]: DEBUG oslo_vmware.api [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5203cf83-60de-07f1-3dbf-08189c2e062a, 'name': SearchDatastore_Task, 'duration_secs': 0.014973} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.179306] env[63531]: DEBUG oslo_concurrency.lockutils [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1121.179573] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] 225e11e5-0558-4ad5-acd4-2a69888fe2d1/225e11e5-0558-4ad5-acd4-2a69888fe2d1.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1121.179834] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-145b639d-4d62-446f-88f5-bbe6dd86ecdd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.190052] env[63531]: DEBUG oslo_vmware.api [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 1121.190052] env[63531]: value = "task-1118689" [ 1121.190052] env[63531]: _type = "Task" [ 1121.190052] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.195350] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0b59a136-b9bd-4427-8763-ab45f0ad4a45 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.203637] env[63531]: DEBUG oslo_vmware.api [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118689, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.211728] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53869991-6315-4d6b-83cf-50c785646d46 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.246809] env[63531]: DEBUG nova.compute.manager [req-6ff58118-8023-4a74-a7cc-c7bdabe51ad8 req-3727232a-2d22-4665-940d-8067d9e1b247 service nova] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] Detach interface failed, port_id=e22c7894-d0a9-4984-993f-0c220a12dfd5, reason: Instance 7fc38de1-1fbf-4312-98f6-6801d5fcd49a could not be found. {{(pid=63531) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1121.324940] env[63531]: INFO nova.compute.manager [-] [instance: 7fc38de1-1fbf-4312-98f6-6801d5fcd49a] Took 1.31 seconds to deallocate network for instance. [ 1121.348466] env[63531]: DEBUG oslo_concurrency.lockutils [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.811s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1121.350712] env[63531]: DEBUG nova.compute.manager [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1121.361565] env[63531]: DEBUG oslo_concurrency.lockutils [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.696s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1121.361565] env[63531]: INFO nova.compute.claims [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1121.602739] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 8f903d83-adbd-4f71-a6f1-037807746248] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1121.670921] env[63531]: DEBUG oslo_vmware.api [None req-7d07d70a-ef6c-4abd-8c0f-95342d2a4723 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118688, 'name': ReconfigVM_Task, 'duration_secs': 0.152184} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.674020] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-7d07d70a-ef6c-4abd-8c0f-95342d2a4723 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244812', 'volume_id': '62b849f1-75ea-47e8-80dd-1e6609042e4a', 'name': 'volume-62b849f1-75ea-47e8-80dd-1e6609042e4a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '89f19669-53a9-4425-a3c6-d5f8873b707e', 'attached_at': '', 'detached_at': '', 'volume_id': '62b849f1-75ea-47e8-80dd-1e6609042e4a', 'serial': '62b849f1-75ea-47e8-80dd-1e6609042e4a'} {{(pid=63531) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1121.701916] env[63531]: DEBUG oslo_vmware.api [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118689, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.451416} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.702416] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] 225e11e5-0558-4ad5-acd4-2a69888fe2d1/225e11e5-0558-4ad5-acd4-2a69888fe2d1.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1121.702522] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1121.702854] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a498b25e-8d52-4b64-bf0b-8be07bd8f43d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.711771] env[63531]: DEBUG oslo_vmware.api [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 1121.711771] env[63531]: value = "task-1118690" [ 1121.711771] env[63531]: _type = "Task" [ 1121.711771] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.722272] env[63531]: DEBUG oslo_vmware.api [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118690, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.831864] env[63531]: DEBUG oslo_concurrency.lockutils [None req-46a89a78-698b-4568-a636-317e94673b85 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1121.867523] env[63531]: DEBUG nova.compute.utils [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1121.873021] env[63531]: DEBUG nova.compute.manager [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1121.873021] env[63531]: DEBUG nova.network.neutron [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1121.924457] env[63531]: DEBUG nova.policy [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2ec68c1860c5484cb8094370baa80c98', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '57e9bf4ee0f1463fa67b0f8deff30fdc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 1122.105758] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 3e7b8a01-d930-43cf-9432-0dabc68da20e] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1122.226575] env[63531]: DEBUG oslo_vmware.api [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118690, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059366} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.227283] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1122.228778] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-809dade6-2537-4d75-9cc1-133afef676f9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.259679] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Reconfiguring VM instance instance-00000071 to attach disk [datastore2] 225e11e5-0558-4ad5-acd4-2a69888fe2d1/225e11e5-0558-4ad5-acd4-2a69888fe2d1.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1122.259679] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8075ed7b-afd9-4380-b0c1-c0bb95597a8c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.281488] env[63531]: DEBUG oslo_vmware.api [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 1122.281488] env[63531]: value = "task-1118691" [ 1122.281488] env[63531]: _type = "Task" [ 1122.281488] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.291039] env[63531]: DEBUG oslo_vmware.api [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118691, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.372973] env[63531]: DEBUG nova.compute.manager [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1122.461487] env[63531]: DEBUG nova.network.neutron [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Successfully created port: 9a3bf9ef-a105-4821-81f8-4891f7c3ae0a {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1122.571061] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a19a6ba8-973b-4c77-b2b3-86812cda9d20 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.579067] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63a3e27b-93b3-46b6-8fa4-6707b8e019eb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.616136] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 97d3f108-e299-4ef1-84ca-532cd64905d5] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1122.619724] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74172293-d217-4c45-a73f-64b9db0a0da7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.630234] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6715f668-77cc-4ff3-b492-c5aaac04c1ee {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.648351] env[63531]: DEBUG nova.compute.provider_tree [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Updating inventory in ProviderTree for provider 9996a314-a62e-4282-8252-f387e7e3fabb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1122.677039] env[63531]: DEBUG nova.compute.manager [req-fa3c1fee-d0be-4c24-aa73-68866a3b8db1 req-5f387691-18b5-4789-b1f5-e03e3806818a service nova] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Received event network-changed-1d65a3d9-a09b-4266-8a05-6abd7ddd18c7 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1122.677260] env[63531]: DEBUG nova.compute.manager [req-fa3c1fee-d0be-4c24-aa73-68866a3b8db1 req-5f387691-18b5-4789-b1f5-e03e3806818a service nova] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Refreshing instance network info cache due to event network-changed-1d65a3d9-a09b-4266-8a05-6abd7ddd18c7. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1122.677499] env[63531]: DEBUG oslo_concurrency.lockutils [req-fa3c1fee-d0be-4c24-aa73-68866a3b8db1 req-5f387691-18b5-4789-b1f5-e03e3806818a service nova] Acquiring lock "refresh_cache-98853767-791e-4568-8bb8-30dfcc473ee3" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1122.677650] env[63531]: DEBUG oslo_concurrency.lockutils [req-fa3c1fee-d0be-4c24-aa73-68866a3b8db1 req-5f387691-18b5-4789-b1f5-e03e3806818a service nova] Acquired lock "refresh_cache-98853767-791e-4568-8bb8-30dfcc473ee3" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1122.677815] env[63531]: DEBUG nova.network.neutron [req-fa3c1fee-d0be-4c24-aa73-68866a3b8db1 req-5f387691-18b5-4789-b1f5-e03e3806818a service nova] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Refreshing network info cache for port 1d65a3d9-a09b-4266-8a05-6abd7ddd18c7 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1122.728664] env[63531]: DEBUG nova.objects.instance [None req-7d07d70a-ef6c-4abd-8c0f-95342d2a4723 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lazy-loading 'flavor' on Instance uuid 89f19669-53a9-4425-a3c6-d5f8873b707e {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1122.781277] env[63531]: DEBUG nova.compute.manager [req-149a204a-8451-4e8a-a6cc-761e683f8f02 req-ad896761-6d1e-4a66-898c-b708c877bcb6 service nova] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Received event network-changed-e9692fb2-f3a9-455c-8317-b533a6af48f7 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1122.781277] env[63531]: DEBUG nova.compute.manager [req-149a204a-8451-4e8a-a6cc-761e683f8f02 req-ad896761-6d1e-4a66-898c-b708c877bcb6 service nova] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Refreshing instance network info cache due to event network-changed-e9692fb2-f3a9-455c-8317-b533a6af48f7. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1122.781277] env[63531]: DEBUG oslo_concurrency.lockutils [req-149a204a-8451-4e8a-a6cc-761e683f8f02 req-ad896761-6d1e-4a66-898c-b708c877bcb6 service nova] Acquiring lock "refresh_cache-ccc8f366-5849-45a6-9745-ff9b2be9c3bb" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1122.781277] env[63531]: DEBUG oslo_concurrency.lockutils [req-149a204a-8451-4e8a-a6cc-761e683f8f02 req-ad896761-6d1e-4a66-898c-b708c877bcb6 service nova] Acquired lock "refresh_cache-ccc8f366-5849-45a6-9745-ff9b2be9c3bb" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1122.781277] env[63531]: DEBUG nova.network.neutron [req-149a204a-8451-4e8a-a6cc-761e683f8f02 req-ad896761-6d1e-4a66-898c-b708c877bcb6 service nova] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Refreshing network info cache for port e9692fb2-f3a9-455c-8317-b533a6af48f7 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1122.796862] env[63531]: DEBUG oslo_vmware.api [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118691, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.003828] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ff6b2803-f13a-4036-94f8-02c56f4fe06e tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquiring lock "89f19669-53a9-4425-a3c6-d5f8873b707e" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.120016] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 96c2910b-1e25-4053-9ad9-c155402ef0ac] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1123.180155] env[63531]: DEBUG nova.scheduler.client.report [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Updated inventory for provider 9996a314-a62e-4282-8252-f387e7e3fabb with generation 158 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1123.180484] env[63531]: DEBUG nova.compute.provider_tree [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Updating resource provider 9996a314-a62e-4282-8252-f387e7e3fabb generation from 158 to 159 during operation: update_inventory {{(pid=63531) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1123.180681] env[63531]: DEBUG nova.compute.provider_tree [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Updating inventory in ProviderTree for provider 9996a314-a62e-4282-8252-f387e7e3fabb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1123.233602] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7d07d70a-ef6c-4abd-8c0f-95342d2a4723 tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "89f19669-53a9-4425-a3c6-d5f8873b707e" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.829s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.234618] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ff6b2803-f13a-4036-94f8-02c56f4fe06e tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "89f19669-53a9-4425-a3c6-d5f8873b707e" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.231s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.297567] env[63531]: DEBUG oslo_vmware.api [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118691, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.394394] env[63531]: DEBUG nova.compute.manager [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1123.428644] env[63531]: DEBUG nova.virt.hardware [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1123.428644] env[63531]: DEBUG nova.virt.hardware [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1123.428644] env[63531]: DEBUG nova.virt.hardware [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1123.428809] env[63531]: DEBUG nova.virt.hardware [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1123.429613] env[63531]: DEBUG nova.virt.hardware [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1123.429613] env[63531]: DEBUG nova.virt.hardware [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1123.429613] env[63531]: DEBUG nova.virt.hardware [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1123.429613] env[63531]: DEBUG nova.virt.hardware [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1123.429830] env[63531]: DEBUG nova.virt.hardware [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1123.429830] env[63531]: DEBUG nova.virt.hardware [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1123.430078] env[63531]: DEBUG nova.virt.hardware [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1123.431796] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d053add-721f-40a1-b48f-6b596d4bb948 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.439902] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09cc95ae-8bd7-4f1f-9a34-42a1e1cf7ce1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.447101] env[63531]: DEBUG nova.network.neutron [req-fa3c1fee-d0be-4c24-aa73-68866a3b8db1 req-5f387691-18b5-4789-b1f5-e03e3806818a service nova] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Updated VIF entry in instance network info cache for port 1d65a3d9-a09b-4266-8a05-6abd7ddd18c7. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1123.447436] env[63531]: DEBUG nova.network.neutron [req-fa3c1fee-d0be-4c24-aa73-68866a3b8db1 req-5f387691-18b5-4789-b1f5-e03e3806818a service nova] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Updating instance_info_cache with network_info: [{"id": "1d65a3d9-a09b-4266-8a05-6abd7ddd18c7", "address": "fa:16:3e:14:b8:48", "network": {"id": "caf86428-59a5-4e50-9135-2ab8473af64d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1673994696-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ba8b284854242c392aec5326e996239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d65a3d9-a0", "ovs_interfaceid": "1d65a3d9-a09b-4266-8a05-6abd7ddd18c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1123.457139] env[63531]: DEBUG oslo_concurrency.lockutils [req-fa3c1fee-d0be-4c24-aa73-68866a3b8db1 req-5f387691-18b5-4789-b1f5-e03e3806818a service nova] Releasing lock "refresh_cache-98853767-791e-4568-8bb8-30dfcc473ee3" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1123.536175] env[63531]: DEBUG nova.network.neutron [req-149a204a-8451-4e8a-a6cc-761e683f8f02 req-ad896761-6d1e-4a66-898c-b708c877bcb6 service nova] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Updated VIF entry in instance network info cache for port e9692fb2-f3a9-455c-8317-b533a6af48f7. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1123.536550] env[63531]: DEBUG nova.network.neutron [req-149a204a-8451-4e8a-a6cc-761e683f8f02 req-ad896761-6d1e-4a66-898c-b708c877bcb6 service nova] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Updating instance_info_cache with network_info: [{"id": "e9692fb2-f3a9-455c-8317-b533a6af48f7", "address": "fa:16:3e:79:4a:3b", "network": {"id": "caf86428-59a5-4e50-9135-2ab8473af64d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1673994696-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ba8b284854242c392aec5326e996239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape9692fb2-f3", "ovs_interfaceid": "e9692fb2-f3a9-455c-8317-b533a6af48f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1123.622956] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: b6db77e6-4447-4b22-93d7-265ff0ada0bd] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1123.688431] env[63531]: DEBUG oslo_concurrency.lockutils [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.330s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.688934] env[63531]: DEBUG nova.compute.manager [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1123.691563] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5af578a2-dd7a-4b46-bfb0-bb8085f7426f tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.136s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.691756] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5af578a2-dd7a-4b46-bfb0-bb8085f7426f tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.693691] env[63531]: DEBUG oslo_concurrency.lockutils [None req-013f893f-db1e-4ad5-ad0b-61a62d7202d5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.077s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.693877] env[63531]: DEBUG oslo_concurrency.lockutils [None req-013f893f-db1e-4ad5-ad0b-61a62d7202d5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.695852] env[63531]: DEBUG oslo_concurrency.lockutils [None req-46a89a78-698b-4568-a636-317e94673b85 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.864s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.696078] env[63531]: DEBUG nova.objects.instance [None req-46a89a78-698b-4568-a636-317e94673b85 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lazy-loading 'resources' on Instance uuid 7fc38de1-1fbf-4312-98f6-6801d5fcd49a {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1123.717097] env[63531]: INFO nova.scheduler.client.report [None req-013f893f-db1e-4ad5-ad0b-61a62d7202d5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Deleted allocations for instance c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9 [ 1123.720023] env[63531]: INFO nova.scheduler.client.report [None req-5af578a2-dd7a-4b46-bfb0-bb8085f7426f tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Deleted allocations for instance 49967454-a4f9-47f6-a2a9-0cd0c29e18d8 [ 1123.738679] env[63531]: INFO nova.compute.manager [None req-ff6b2803-f13a-4036-94f8-02c56f4fe06e tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Detaching volume 62b849f1-75ea-47e8-80dd-1e6609042e4a [ 1123.783363] env[63531]: INFO nova.virt.block_device [None req-ff6b2803-f13a-4036-94f8-02c56f4fe06e tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Attempting to driver detach volume 62b849f1-75ea-47e8-80dd-1e6609042e4a from mountpoint /dev/sdb [ 1123.783667] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-ff6b2803-f13a-4036-94f8-02c56f4fe06e tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Volume detach. Driver type: vmdk {{(pid=63531) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1123.784044] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-ff6b2803-f13a-4036-94f8-02c56f4fe06e tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244812', 'volume_id': '62b849f1-75ea-47e8-80dd-1e6609042e4a', 'name': 'volume-62b849f1-75ea-47e8-80dd-1e6609042e4a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '89f19669-53a9-4425-a3c6-d5f8873b707e', 'attached_at': '', 'detached_at': '', 'volume_id': '62b849f1-75ea-47e8-80dd-1e6609042e4a', 'serial': '62b849f1-75ea-47e8-80dd-1e6609042e4a'} {{(pid=63531) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1123.784811] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae6482b2-c652-4426-add9-c8543348b10e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.797037] env[63531]: DEBUG oslo_vmware.api [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118691, 'name': ReconfigVM_Task, 'duration_secs': 1.0683} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.817080] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Reconfigured VM instance instance-00000071 to attach disk [datastore2] 225e11e5-0558-4ad5-acd4-2a69888fe2d1/225e11e5-0558-4ad5-acd4-2a69888fe2d1.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1123.817080] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0459ac6c-8d9d-4837-9042-d283c82120fa {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.817080] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfbc1f62-a0f0-4b38-9015-8ec3154b8aa1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.824443] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45bd8b90-c02a-4387-89b3-8a3eef6a49d3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.827799] env[63531]: DEBUG oslo_vmware.api [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 1123.827799] env[63531]: value = "task-1118692" [ 1123.827799] env[63531]: _type = "Task" [ 1123.827799] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.848292] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cbff5e9-16b1-4b50-ba3e-de707c3870f7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.854158] env[63531]: DEBUG oslo_vmware.api [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118692, 'name': Rename_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.868466] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-ff6b2803-f13a-4036-94f8-02c56f4fe06e tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] The volume has not been displaced from its original location: [datastore2] volume-62b849f1-75ea-47e8-80dd-1e6609042e4a/volume-62b849f1-75ea-47e8-80dd-1e6609042e4a.vmdk. No consolidation needed. {{(pid=63531) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1123.873295] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-ff6b2803-f13a-4036-94f8-02c56f4fe06e tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Reconfiguring VM instance instance-00000069 to detach disk 2001 {{(pid=63531) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1123.873606] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1fa1af07-4c80-416c-b026-f10d22b22c72 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.894066] env[63531]: DEBUG oslo_vmware.api [None req-ff6b2803-f13a-4036-94f8-02c56f4fe06e tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for the task: (returnval){ [ 1123.894066] env[63531]: value = "task-1118693" [ 1123.894066] env[63531]: _type = "Task" [ 1123.894066] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.905129] env[63531]: DEBUG oslo_vmware.api [None req-ff6b2803-f13a-4036-94f8-02c56f4fe06e tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118693, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.039780] env[63531]: DEBUG oslo_concurrency.lockutils [req-149a204a-8451-4e8a-a6cc-761e683f8f02 req-ad896761-6d1e-4a66-898c-b708c877bcb6 service nova] Releasing lock "refresh_cache-ccc8f366-5849-45a6-9745-ff9b2be9c3bb" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1124.040237] env[63531]: DEBUG nova.compute.manager [req-149a204a-8451-4e8a-a6cc-761e683f8f02 req-ad896761-6d1e-4a66-898c-b708c877bcb6 service nova] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Received event network-changed-1d65a3d9-a09b-4266-8a05-6abd7ddd18c7 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1124.040435] env[63531]: DEBUG nova.compute.manager [req-149a204a-8451-4e8a-a6cc-761e683f8f02 req-ad896761-6d1e-4a66-898c-b708c877bcb6 service nova] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Refreshing instance network info cache due to event network-changed-1d65a3d9-a09b-4266-8a05-6abd7ddd18c7. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1124.040671] env[63531]: DEBUG oslo_concurrency.lockutils [req-149a204a-8451-4e8a-a6cc-761e683f8f02 req-ad896761-6d1e-4a66-898c-b708c877bcb6 service nova] Acquiring lock "refresh_cache-98853767-791e-4568-8bb8-30dfcc473ee3" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1124.040871] env[63531]: DEBUG oslo_concurrency.lockutils [req-149a204a-8451-4e8a-a6cc-761e683f8f02 req-ad896761-6d1e-4a66-898c-b708c877bcb6 service nova] Acquired lock "refresh_cache-98853767-791e-4568-8bb8-30dfcc473ee3" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1124.041343] env[63531]: DEBUG nova.network.neutron [req-149a204a-8451-4e8a-a6cc-761e683f8f02 req-ad896761-6d1e-4a66-898c-b708c877bcb6 service nova] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Refreshing network info cache for port 1d65a3d9-a09b-4266-8a05-6abd7ddd18c7 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1124.126033] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: ba97ce22-ea4f-430b-a4c6-08204a568436] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1124.166577] env[63531]: DEBUG nova.network.neutron [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Successfully updated port: 9a3bf9ef-a105-4821-81f8-4891f7c3ae0a {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1124.200063] env[63531]: DEBUG nova.compute.utils [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1124.203719] env[63531]: DEBUG nova.compute.manager [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1124.203885] env[63531]: DEBUG nova.network.neutron [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1124.233196] env[63531]: DEBUG oslo_concurrency.lockutils [None req-013f893f-db1e-4ad5-ad0b-61a62d7202d5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "c80ea5ae-ee1f-4d6d-a29f-24f9c98ecae9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.554s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1124.234170] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5af578a2-dd7a-4b46-bfb0-bb8085f7426f tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "49967454-a4f9-47f6-a2a9-0cd0c29e18d8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.403s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1124.234920] env[63531]: DEBUG oslo_concurrency.lockutils [None req-99659a4d-3f40-428c-9238-204dfed9bdfb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "49967454-a4f9-47f6-a2a9-0cd0c29e18d8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 12.090s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1124.235158] env[63531]: DEBUG oslo_concurrency.lockutils [None req-99659a4d-3f40-428c-9238-204dfed9bdfb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquiring lock "49967454-a4f9-47f6-a2a9-0cd0c29e18d8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1124.235367] env[63531]: DEBUG oslo_concurrency.lockutils [None req-99659a4d-3f40-428c-9238-204dfed9bdfb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "49967454-a4f9-47f6-a2a9-0cd0c29e18d8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1124.235536] env[63531]: DEBUG oslo_concurrency.lockutils [None req-99659a4d-3f40-428c-9238-204dfed9bdfb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "49967454-a4f9-47f6-a2a9-0cd0c29e18d8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1124.241191] env[63531]: INFO nova.compute.manager [None req-99659a4d-3f40-428c-9238-204dfed9bdfb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Terminating instance [ 1124.242579] env[63531]: DEBUG nova.compute.manager [None req-99659a4d-3f40-428c-9238-204dfed9bdfb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1124.242963] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-eb3be430-7b9d-403f-81c9-bc46ef1604fb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.246856] env[63531]: DEBUG nova.policy [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '597d424341124b7db23dc7a104107148', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '308eabafea9c4188a58a0f1c22074d2f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 1124.260484] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17aa2527-97cb-455b-948c-a85d2bbb4797 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.296497] env[63531]: WARNING nova.virt.vmwareapi.driver [None req-99659a4d-3f40-428c-9238-204dfed9bdfb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 49967454-a4f9-47f6-a2a9-0cd0c29e18d8 could not be found. [ 1124.296729] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-99659a4d-3f40-428c-9238-204dfed9bdfb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1124.300751] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ddab401d-9d55-4fef-838b-d18ccc569791 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.310439] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d326f6d-4a6d-4156-ba3f-cd4bf836d126 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.341411] env[63531]: WARNING nova.virt.vmwareapi.vmops [None req-99659a4d-3f40-428c-9238-204dfed9bdfb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 49967454-a4f9-47f6-a2a9-0cd0c29e18d8 could not be found. [ 1124.341637] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-99659a4d-3f40-428c-9238-204dfed9bdfb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1124.341817] env[63531]: INFO nova.compute.manager [None req-99659a4d-3f40-428c-9238-204dfed9bdfb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Took 0.10 seconds to destroy the instance on the hypervisor. [ 1124.342110] env[63531]: DEBUG oslo.service.loopingcall [None req-99659a4d-3f40-428c-9238-204dfed9bdfb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1124.347173] env[63531]: DEBUG nova.compute.manager [-] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1124.348222] env[63531]: DEBUG nova.network.neutron [-] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1124.349741] env[63531]: DEBUG oslo_vmware.api [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118692, 'name': Rename_Task, 'duration_secs': 0.23991} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.350219] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1124.350712] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f965bfb3-f71d-4e9d-ac82-ccb9e1e534f7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.361590] env[63531]: DEBUG oslo_vmware.api [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 1124.361590] env[63531]: value = "task-1118695" [ 1124.361590] env[63531]: _type = "Task" [ 1124.361590] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.372726] env[63531]: DEBUG oslo_vmware.api [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118695, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.379056] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db3e0e52-810f-44ec-9bd5-0ccd5c7d5c49 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.386953] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38090142-478d-421c-8863-c792d6b44b9c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.425320] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-567a0231-8a06-459a-a9a0-cb4198ceb0b1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.430968] env[63531]: DEBUG oslo_vmware.api [None req-ff6b2803-f13a-4036-94f8-02c56f4fe06e tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118693, 'name': ReconfigVM_Task, 'duration_secs': 0.237353} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.432074] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-ff6b2803-f13a-4036-94f8-02c56f4fe06e tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Reconfigured VM instance instance-00000069 to detach disk 2001 {{(pid=63531) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1124.438194] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c87438b4-210e-43c5-a9da-a9cbaf23be2f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.449906] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-933897f7-6502-4307-9c36-a8cf1b7454a1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.465978] env[63531]: DEBUG nova.compute.provider_tree [None req-46a89a78-698b-4568-a636-317e94673b85 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1124.468524] env[63531]: DEBUG oslo_vmware.api [None req-ff6b2803-f13a-4036-94f8-02c56f4fe06e tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for the task: (returnval){ [ 1124.468524] env[63531]: value = "task-1118696" [ 1124.468524] env[63531]: _type = "Task" [ 1124.468524] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.478432] env[63531]: DEBUG oslo_vmware.api [None req-ff6b2803-f13a-4036-94f8-02c56f4fe06e tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118696, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.556041] env[63531]: DEBUG nova.network.neutron [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Successfully created port: 8c2307f2-c72e-47bb-a1f7-1dafecb0d2a4 {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1124.628775] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: afa5f2c8-e519-4eb7-8e0e-ada9efb630e7] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1124.670635] env[63531]: DEBUG oslo_concurrency.lockutils [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquiring lock "refresh_cache-3794f44c-8f6d-4bab-b9e0-9f6c819e1a97" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1124.670635] env[63531]: DEBUG oslo_concurrency.lockutils [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquired lock "refresh_cache-3794f44c-8f6d-4bab-b9e0-9f6c819e1a97" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1124.670635] env[63531]: DEBUG nova.network.neutron [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1124.701916] env[63531]: DEBUG nova.compute.manager [req-d2e5bf37-b25d-4fe8-a9da-e814e2189291 req-1b982a2b-f852-4a34-964e-81d49be1b415 service nova] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Received event network-changed-e9692fb2-f3a9-455c-8317-b533a6af48f7 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1124.702124] env[63531]: DEBUG nova.compute.manager [req-d2e5bf37-b25d-4fe8-a9da-e814e2189291 req-1b982a2b-f852-4a34-964e-81d49be1b415 service nova] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Refreshing instance network info cache due to event network-changed-e9692fb2-f3a9-455c-8317-b533a6af48f7. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1124.702339] env[63531]: DEBUG oslo_concurrency.lockutils [req-d2e5bf37-b25d-4fe8-a9da-e814e2189291 req-1b982a2b-f852-4a34-964e-81d49be1b415 service nova] Acquiring lock "refresh_cache-ccc8f366-5849-45a6-9745-ff9b2be9c3bb" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1124.702499] env[63531]: DEBUG oslo_concurrency.lockutils [req-d2e5bf37-b25d-4fe8-a9da-e814e2189291 req-1b982a2b-f852-4a34-964e-81d49be1b415 service nova] Acquired lock "refresh_cache-ccc8f366-5849-45a6-9745-ff9b2be9c3bb" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1124.702670] env[63531]: DEBUG nova.network.neutron [req-d2e5bf37-b25d-4fe8-a9da-e814e2189291 req-1b982a2b-f852-4a34-964e-81d49be1b415 service nova] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Refreshing network info cache for port e9692fb2-f3a9-455c-8317-b533a6af48f7 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1124.704195] env[63531]: DEBUG nova.compute.manager [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1124.769515] env[63531]: DEBUG nova.network.neutron [req-149a204a-8451-4e8a-a6cc-761e683f8f02 req-ad896761-6d1e-4a66-898c-b708c877bcb6 service nova] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Updated VIF entry in instance network info cache for port 1d65a3d9-a09b-4266-8a05-6abd7ddd18c7. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1124.769906] env[63531]: DEBUG nova.network.neutron [req-149a204a-8451-4e8a-a6cc-761e683f8f02 req-ad896761-6d1e-4a66-898c-b708c877bcb6 service nova] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Updating instance_info_cache with network_info: [{"id": "1d65a3d9-a09b-4266-8a05-6abd7ddd18c7", "address": "fa:16:3e:14:b8:48", "network": {"id": "caf86428-59a5-4e50-9135-2ab8473af64d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1673994696-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ba8b284854242c392aec5326e996239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d65a3d9-a0", "ovs_interfaceid": "1d65a3d9-a09b-4266-8a05-6abd7ddd18c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1124.806736] env[63531]: DEBUG nova.compute.manager [req-80adf1c9-75c5-454c-8f7c-8990d48c8d9f req-4a7b46c1-bb9e-496f-9a78-248c2a974495 service nova] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Received event network-vif-plugged-9a3bf9ef-a105-4821-81f8-4891f7c3ae0a {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1124.806949] env[63531]: DEBUG oslo_concurrency.lockutils [req-80adf1c9-75c5-454c-8f7c-8990d48c8d9f req-4a7b46c1-bb9e-496f-9a78-248c2a974495 service nova] Acquiring lock "3794f44c-8f6d-4bab-b9e0-9f6c819e1a97-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1124.807211] env[63531]: DEBUG oslo_concurrency.lockutils [req-80adf1c9-75c5-454c-8f7c-8990d48c8d9f req-4a7b46c1-bb9e-496f-9a78-248c2a974495 service nova] Lock "3794f44c-8f6d-4bab-b9e0-9f6c819e1a97-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1124.807399] env[63531]: DEBUG oslo_concurrency.lockutils [req-80adf1c9-75c5-454c-8f7c-8990d48c8d9f req-4a7b46c1-bb9e-496f-9a78-248c2a974495 service nova] Lock "3794f44c-8f6d-4bab-b9e0-9f6c819e1a97-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1124.807576] env[63531]: DEBUG nova.compute.manager [req-80adf1c9-75c5-454c-8f7c-8990d48c8d9f req-4a7b46c1-bb9e-496f-9a78-248c2a974495 service nova] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] No waiting events found dispatching network-vif-plugged-9a3bf9ef-a105-4821-81f8-4891f7c3ae0a {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1124.807746] env[63531]: WARNING nova.compute.manager [req-80adf1c9-75c5-454c-8f7c-8990d48c8d9f req-4a7b46c1-bb9e-496f-9a78-248c2a974495 service nova] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Received unexpected event network-vif-plugged-9a3bf9ef-a105-4821-81f8-4891f7c3ae0a for instance with vm_state building and task_state spawning. [ 1124.808252] env[63531]: DEBUG nova.compute.manager [req-80adf1c9-75c5-454c-8f7c-8990d48c8d9f req-4a7b46c1-bb9e-496f-9a78-248c2a974495 service nova] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Received event network-changed-9a3bf9ef-a105-4821-81f8-4891f7c3ae0a {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1124.808252] env[63531]: DEBUG nova.compute.manager [req-80adf1c9-75c5-454c-8f7c-8990d48c8d9f req-4a7b46c1-bb9e-496f-9a78-248c2a974495 service nova] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Refreshing instance network info cache due to event network-changed-9a3bf9ef-a105-4821-81f8-4891f7c3ae0a. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1124.808252] env[63531]: DEBUG oslo_concurrency.lockutils [req-80adf1c9-75c5-454c-8f7c-8990d48c8d9f req-4a7b46c1-bb9e-496f-9a78-248c2a974495 service nova] Acquiring lock "refresh_cache-3794f44c-8f6d-4bab-b9e0-9f6c819e1a97" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1124.868747] env[63531]: DEBUG oslo_vmware.api [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118695, 'name': PowerOnVM_Task} progress is 81%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.874136] env[63531]: DEBUG nova.network.neutron [-] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1124.969932] env[63531]: DEBUG nova.scheduler.client.report [None req-46a89a78-698b-4568-a636-317e94673b85 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1124.982377] env[63531]: DEBUG oslo_vmware.api [None req-ff6b2803-f13a-4036-94f8-02c56f4fe06e tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118696, 'name': ReconfigVM_Task, 'duration_secs': 0.155127} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.983252] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-ff6b2803-f13a-4036-94f8-02c56f4fe06e tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-244812', 'volume_id': '62b849f1-75ea-47e8-80dd-1e6609042e4a', 'name': 'volume-62b849f1-75ea-47e8-80dd-1e6609042e4a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '89f19669-53a9-4425-a3c6-d5f8873b707e', 'attached_at': '', 'detached_at': '', 'volume_id': '62b849f1-75ea-47e8-80dd-1e6609042e4a', 'serial': '62b849f1-75ea-47e8-80dd-1e6609042e4a'} {{(pid=63531) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1125.134754] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 5f91a7da-ca2f-42ab-8cc0-c6b0fac6e5a7] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1125.270567] env[63531]: DEBUG nova.network.neutron [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1125.273265] env[63531]: DEBUG oslo_concurrency.lockutils [req-149a204a-8451-4e8a-a6cc-761e683f8f02 req-ad896761-6d1e-4a66-898c-b708c877bcb6 service nova] Releasing lock "refresh_cache-98853767-791e-4568-8bb8-30dfcc473ee3" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1125.370680] env[63531]: DEBUG oslo_vmware.api [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118695, 'name': PowerOnVM_Task, 'duration_secs': 0.885262} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.370984] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1125.371210] env[63531]: INFO nova.compute.manager [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Took 8.37 seconds to spawn the instance on the hypervisor. [ 1125.371521] env[63531]: DEBUG nova.compute.manager [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1125.375186] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c60a2da-b993-4078-9c7a-fbb27d8c3aa1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.378800] env[63531]: INFO nova.compute.manager [-] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Took 1.03 seconds to deallocate network for instance. [ 1125.394185] env[63531]: WARNING nova.volume.cinder [None req-99659a4d-3f40-428c-9238-204dfed9bdfb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Attachment 4db59766-5c8c-4c09-bb08-c72e703547ba does not exist. Ignoring.: cinderclient.exceptions.NotFound: Volume attachment could not be found with filter: attachment_id = 4db59766-5c8c-4c09-bb08-c72e703547ba. (HTTP 404) (Request-ID: req-9f390eb6-cd9c-4636-98ed-766d853b0fa7) [ 1125.394455] env[63531]: INFO nova.compute.manager [None req-99659a4d-3f40-428c-9238-204dfed9bdfb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Took 0.02 seconds to detach 1 volumes for instance. [ 1125.396382] env[63531]: DEBUG nova.compute.manager [None req-99659a4d-3f40-428c-9238-204dfed9bdfb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Deleting volume: 422ea428-1dc4-40f2-8151-a5e18a596f9b {{(pid=63531) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 1125.409815] env[63531]: WARNING nova.compute.manager [None req-99659a4d-3f40-428c-9238-204dfed9bdfb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Failed to delete volume: 422ea428-1dc4-40f2-8151-a5e18a596f9b due to Volume 422ea428-1dc4-40f2-8151-a5e18a596f9b could not be found.: nova.exception.VolumeNotFound: Volume 422ea428-1dc4-40f2-8151-a5e18a596f9b could not be found. [ 1125.451923] env[63531]: DEBUG nova.network.neutron [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Updating instance_info_cache with network_info: [{"id": "9a3bf9ef-a105-4821-81f8-4891f7c3ae0a", "address": "fa:16:3e:dd:5e:bd", "network": {"id": "c4f4c994-36ab-42ab-94f8-78f29cbafa50", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-994507698-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "57e9bf4ee0f1463fa67b0f8deff30fdc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "767a3a48-41d4-4a0c-961d-0024837f63bd", "external-id": "nsx-vlan-transportzone-141", "segmentation_id": 141, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a3bf9ef-a1", "ovs_interfaceid": "9a3bf9ef-a105-4821-81f8-4891f7c3ae0a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1125.458866] env[63531]: DEBUG nova.network.neutron [req-d2e5bf37-b25d-4fe8-a9da-e814e2189291 req-1b982a2b-f852-4a34-964e-81d49be1b415 service nova] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Updated VIF entry in instance network info cache for port e9692fb2-f3a9-455c-8317-b533a6af48f7. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1125.459221] env[63531]: DEBUG nova.network.neutron [req-d2e5bf37-b25d-4fe8-a9da-e814e2189291 req-1b982a2b-f852-4a34-964e-81d49be1b415 service nova] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Updating instance_info_cache with network_info: [{"id": "e9692fb2-f3a9-455c-8317-b533a6af48f7", "address": "fa:16:3e:79:4a:3b", "network": {"id": "caf86428-59a5-4e50-9135-2ab8473af64d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1673994696-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ba8b284854242c392aec5326e996239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape9692fb2-f3", "ovs_interfaceid": "e9692fb2-f3a9-455c-8317-b533a6af48f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1125.477698] env[63531]: DEBUG oslo_concurrency.lockutils [None req-46a89a78-698b-4568-a636-317e94673b85 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.782s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1125.498995] env[63531]: INFO nova.scheduler.client.report [None req-46a89a78-698b-4568-a636-317e94673b85 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Deleted allocations for instance 7fc38de1-1fbf-4312-98f6-6801d5fcd49a [ 1125.528602] env[63531]: DEBUG nova.objects.instance [None req-ff6b2803-f13a-4036-94f8-02c56f4fe06e tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lazy-loading 'flavor' on Instance uuid 89f19669-53a9-4425-a3c6-d5f8873b707e {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1125.638378] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: b7b2b7ac-2321-4cf3-b2ed-4877b893af63] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1125.715822] env[63531]: DEBUG nova.compute.manager [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1125.743564] env[63531]: DEBUG nova.virt.hardware [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1125.743814] env[63531]: DEBUG nova.virt.hardware [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1125.743980] env[63531]: DEBUG nova.virt.hardware [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1125.744179] env[63531]: DEBUG nova.virt.hardware [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1125.744332] env[63531]: DEBUG nova.virt.hardware [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1125.744484] env[63531]: DEBUG nova.virt.hardware [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1125.744696] env[63531]: DEBUG nova.virt.hardware [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1125.744860] env[63531]: DEBUG nova.virt.hardware [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1125.745042] env[63531]: DEBUG nova.virt.hardware [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1125.745230] env[63531]: DEBUG nova.virt.hardware [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1125.745415] env[63531]: DEBUG nova.virt.hardware [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1125.746320] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-788d8b7c-d43b-4788-bef7-e91f3b579c3a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.754977] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2665482-18e0-40bb-963b-9fb4bdd8e8fc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.899319] env[63531]: INFO nova.compute.manager [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Took 15.00 seconds to build instance. [ 1125.914997] env[63531]: INFO nova.compute.manager [None req-99659a4d-3f40-428c-9238-204dfed9bdfb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 49967454-a4f9-47f6-a2a9-0cd0c29e18d8] Instance disappeared during terminate [ 1125.915271] env[63531]: DEBUG oslo_concurrency.lockutils [None req-99659a4d-3f40-428c-9238-204dfed9bdfb tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "49967454-a4f9-47f6-a2a9-0cd0c29e18d8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 1.680s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1125.954595] env[63531]: DEBUG oslo_concurrency.lockutils [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Releasing lock "refresh_cache-3794f44c-8f6d-4bab-b9e0-9f6c819e1a97" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1125.954894] env[63531]: DEBUG nova.compute.manager [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Instance network_info: |[{"id": "9a3bf9ef-a105-4821-81f8-4891f7c3ae0a", "address": "fa:16:3e:dd:5e:bd", "network": {"id": "c4f4c994-36ab-42ab-94f8-78f29cbafa50", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-994507698-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "57e9bf4ee0f1463fa67b0f8deff30fdc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "767a3a48-41d4-4a0c-961d-0024837f63bd", "external-id": "nsx-vlan-transportzone-141", "segmentation_id": 141, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a3bf9ef-a1", "ovs_interfaceid": "9a3bf9ef-a105-4821-81f8-4891f7c3ae0a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1125.955242] env[63531]: DEBUG oslo_concurrency.lockutils [req-80adf1c9-75c5-454c-8f7c-8990d48c8d9f req-4a7b46c1-bb9e-496f-9a78-248c2a974495 service nova] Acquired lock "refresh_cache-3794f44c-8f6d-4bab-b9e0-9f6c819e1a97" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1125.955441] env[63531]: DEBUG nova.network.neutron [req-80adf1c9-75c5-454c-8f7c-8990d48c8d9f req-4a7b46c1-bb9e-496f-9a78-248c2a974495 service nova] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Refreshing network info cache for port 9a3bf9ef-a105-4821-81f8-4891f7c3ae0a {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1125.956611] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:dd:5e:bd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '767a3a48-41d4-4a0c-961d-0024837f63bd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9a3bf9ef-a105-4821-81f8-4891f7c3ae0a', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1125.964387] env[63531]: DEBUG oslo.service.loopingcall [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1125.965997] env[63531]: DEBUG oslo_concurrency.lockutils [req-d2e5bf37-b25d-4fe8-a9da-e814e2189291 req-1b982a2b-f852-4a34-964e-81d49be1b415 service nova] Releasing lock "refresh_cache-ccc8f366-5849-45a6-9745-ff9b2be9c3bb" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1125.966372] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1125.966728] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-672a6dce-88ba-47a8-86dd-ab5e5cbee1ef {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.988769] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1125.988769] env[63531]: value = "task-1118697" [ 1125.988769] env[63531]: _type = "Task" [ 1125.988769] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.997988] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118697, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.010409] env[63531]: DEBUG oslo_concurrency.lockutils [None req-46a89a78-698b-4568-a636-317e94673b85 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "7fc38de1-1fbf-4312-98f6-6801d5fcd49a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.628s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1126.054925] env[63531]: DEBUG nova.network.neutron [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Successfully updated port: 8c2307f2-c72e-47bb-a1f7-1dafecb0d2a4 {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1126.141858] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 18057005-1ff1-4e3d-bccc-8c89755d6db7] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1126.401692] env[63531]: DEBUG oslo_concurrency.lockutils [None req-00c89b7a-e5af-4492-9b4e-3add4a9d5ab0 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "225e11e5-0558-4ad5-acd4-2a69888fe2d1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.507s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1126.499192] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118697, 'name': CreateVM_Task, 'duration_secs': 0.32699} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.499378] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1126.500070] env[63531]: DEBUG oslo_concurrency.lockutils [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1126.500351] env[63531]: DEBUG oslo_concurrency.lockutils [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1126.500694] env[63531]: DEBUG oslo_concurrency.lockutils [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1126.500965] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-be178a85-4465-4b6d-82f6-b7678082b03e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.506660] env[63531]: DEBUG oslo_vmware.api [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 1126.506660] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52597146-985d-e308-4d32-af10d3f6bea0" [ 1126.506660] env[63531]: _type = "Task" [ 1126.506660] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.516074] env[63531]: DEBUG oslo_vmware.api [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52597146-985d-e308-4d32-af10d3f6bea0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.536760] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ff6b2803-f13a-4036-94f8-02c56f4fe06e tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "89f19669-53a9-4425-a3c6-d5f8873b707e" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.302s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1126.558293] env[63531]: DEBUG oslo_concurrency.lockutils [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "refresh_cache-2d51997d-f88c-468c-9a56-0331a06ca56c" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1126.558515] env[63531]: DEBUG oslo_concurrency.lockutils [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquired lock "refresh_cache-2d51997d-f88c-468c-9a56-0331a06ca56c" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1126.558720] env[63531]: DEBUG nova.network.neutron [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1126.645652] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: e742d3af-5a45-4274-adfb-c6138bf84d6a] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1126.674074] env[63531]: DEBUG nova.network.neutron [req-80adf1c9-75c5-454c-8f7c-8990d48c8d9f req-4a7b46c1-bb9e-496f-9a78-248c2a974495 service nova] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Updated VIF entry in instance network info cache for port 9a3bf9ef-a105-4821-81f8-4891f7c3ae0a. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1126.674530] env[63531]: DEBUG nova.network.neutron [req-80adf1c9-75c5-454c-8f7c-8990d48c8d9f req-4a7b46c1-bb9e-496f-9a78-248c2a974495 service nova] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Updating instance_info_cache with network_info: [{"id": "9a3bf9ef-a105-4821-81f8-4891f7c3ae0a", "address": "fa:16:3e:dd:5e:bd", "network": {"id": "c4f4c994-36ab-42ab-94f8-78f29cbafa50", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-994507698-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "57e9bf4ee0f1463fa67b0f8deff30fdc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "767a3a48-41d4-4a0c-961d-0024837f63bd", "external-id": "nsx-vlan-transportzone-141", "segmentation_id": 141, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a3bf9ef-a1", "ovs_interfaceid": "9a3bf9ef-a105-4821-81f8-4891f7c3ae0a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1126.833755] env[63531]: DEBUG nova.compute.manager [req-ae346bcf-d1a4-4755-89b1-48f3545c8e02 req-9f2f97cc-5b51-4939-b4f1-51203fdddfde service nova] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Received event network-vif-plugged-8c2307f2-c72e-47bb-a1f7-1dafecb0d2a4 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1126.833988] env[63531]: DEBUG oslo_concurrency.lockutils [req-ae346bcf-d1a4-4755-89b1-48f3545c8e02 req-9f2f97cc-5b51-4939-b4f1-51203fdddfde service nova] Acquiring lock "2d51997d-f88c-468c-9a56-0331a06ca56c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1126.834402] env[63531]: DEBUG oslo_concurrency.lockutils [req-ae346bcf-d1a4-4755-89b1-48f3545c8e02 req-9f2f97cc-5b51-4939-b4f1-51203fdddfde service nova] Lock "2d51997d-f88c-468c-9a56-0331a06ca56c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1126.834601] env[63531]: DEBUG oslo_concurrency.lockutils [req-ae346bcf-d1a4-4755-89b1-48f3545c8e02 req-9f2f97cc-5b51-4939-b4f1-51203fdddfde service nova] Lock "2d51997d-f88c-468c-9a56-0331a06ca56c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1126.834885] env[63531]: DEBUG nova.compute.manager [req-ae346bcf-d1a4-4755-89b1-48f3545c8e02 req-9f2f97cc-5b51-4939-b4f1-51203fdddfde service nova] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] No waiting events found dispatching network-vif-plugged-8c2307f2-c72e-47bb-a1f7-1dafecb0d2a4 {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1126.835248] env[63531]: WARNING nova.compute.manager [req-ae346bcf-d1a4-4755-89b1-48f3545c8e02 req-9f2f97cc-5b51-4939-b4f1-51203fdddfde service nova] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Received unexpected event network-vif-plugged-8c2307f2-c72e-47bb-a1f7-1dafecb0d2a4 for instance with vm_state building and task_state spawning. [ 1126.835541] env[63531]: DEBUG nova.compute.manager [req-ae346bcf-d1a4-4755-89b1-48f3545c8e02 req-9f2f97cc-5b51-4939-b4f1-51203fdddfde service nova] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Received event network-changed-8c2307f2-c72e-47bb-a1f7-1dafecb0d2a4 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1126.835812] env[63531]: DEBUG nova.compute.manager [req-ae346bcf-d1a4-4755-89b1-48f3545c8e02 req-9f2f97cc-5b51-4939-b4f1-51203fdddfde service nova] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Refreshing instance network info cache due to event network-changed-8c2307f2-c72e-47bb-a1f7-1dafecb0d2a4. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1126.836090] env[63531]: DEBUG oslo_concurrency.lockutils [req-ae346bcf-d1a4-4755-89b1-48f3545c8e02 req-9f2f97cc-5b51-4939-b4f1-51203fdddfde service nova] Acquiring lock "refresh_cache-2d51997d-f88c-468c-9a56-0331a06ca56c" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1127.019592] env[63531]: DEBUG oslo_vmware.api [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52597146-985d-e308-4d32-af10d3f6bea0, 'name': SearchDatastore_Task, 'duration_secs': 0.009971} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.019900] env[63531]: DEBUG oslo_concurrency.lockutils [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1127.020214] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1127.020438] env[63531]: DEBUG oslo_concurrency.lockutils [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1127.020594] env[63531]: DEBUG oslo_concurrency.lockutils [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1127.020779] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1127.021055] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a02ba0b4-564f-4c84-ab52-8a5691d4e557 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.030208] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1127.030389] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1127.031124] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-37929bf4-eea4-499d-9f32-da82e2f46edd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.036933] env[63531]: DEBUG oslo_vmware.api [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 1127.036933] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]526a6a0d-cdf8-35e7-4e30-e4e06f418b69" [ 1127.036933] env[63531]: _type = "Task" [ 1127.036933] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.046091] env[63531]: DEBUG oslo_vmware.api [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]526a6a0d-cdf8-35e7-4e30-e4e06f418b69, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.101393] env[63531]: DEBUG nova.network.neutron [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1127.149367] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 4757625a-8398-4b2d-b6fe-40eea8913068] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1127.176659] env[63531]: DEBUG oslo_concurrency.lockutils [req-80adf1c9-75c5-454c-8f7c-8990d48c8d9f req-4a7b46c1-bb9e-496f-9a78-248c2a974495 service nova] Releasing lock "refresh_cache-3794f44c-8f6d-4bab-b9e0-9f6c819e1a97" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1127.254468] env[63531]: DEBUG nova.network.neutron [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Updating instance_info_cache with network_info: [{"id": "8c2307f2-c72e-47bb-a1f7-1dafecb0d2a4", "address": "fa:16:3e:54:75:66", "network": {"id": "0667c176-cfdf-4e43-b152-e00bdc224448", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-243715985-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "308eabafea9c4188a58a0f1c22074d2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8f580e6-1d86-41ee-9ebe-c531cb9299c6", "external-id": "nsx-vlan-transportzone-150", "segmentation_id": 150, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c2307f2-c7", "ovs_interfaceid": "8c2307f2-c72e-47bb-a1f7-1dafecb0d2a4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1127.295513] env[63531]: DEBUG oslo_concurrency.lockutils [None req-26695b07-87c6-42d0-9fcc-9d1cd72db1af tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquiring lock "89f19669-53a9-4425-a3c6-d5f8873b707e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1127.295794] env[63531]: DEBUG oslo_concurrency.lockutils [None req-26695b07-87c6-42d0-9fcc-9d1cd72db1af tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "89f19669-53a9-4425-a3c6-d5f8873b707e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1127.296030] env[63531]: DEBUG oslo_concurrency.lockutils [None req-26695b07-87c6-42d0-9fcc-9d1cd72db1af tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquiring lock "89f19669-53a9-4425-a3c6-d5f8873b707e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1127.296227] env[63531]: DEBUG oslo_concurrency.lockutils [None req-26695b07-87c6-42d0-9fcc-9d1cd72db1af tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "89f19669-53a9-4425-a3c6-d5f8873b707e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1127.296404] env[63531]: DEBUG oslo_concurrency.lockutils [None req-26695b07-87c6-42d0-9fcc-9d1cd72db1af tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "89f19669-53a9-4425-a3c6-d5f8873b707e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1127.299037] env[63531]: INFO nova.compute.manager [None req-26695b07-87c6-42d0-9fcc-9d1cd72db1af tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Terminating instance [ 1127.301107] env[63531]: DEBUG nova.compute.manager [None req-26695b07-87c6-42d0-9fcc-9d1cd72db1af tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1127.301413] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-26695b07-87c6-42d0-9fcc-9d1cd72db1af tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1127.302559] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b716616-5aea-42fa-b427-1ce76d31a7c6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.310987] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-26695b07-87c6-42d0-9fcc-9d1cd72db1af tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1127.311260] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9b1d4ffd-0c0a-484f-92f2-06dcd23e99bb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.324982] env[63531]: DEBUG oslo_vmware.api [None req-26695b07-87c6-42d0-9fcc-9d1cd72db1af tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for the task: (returnval){ [ 1127.324982] env[63531]: value = "task-1118698" [ 1127.324982] env[63531]: _type = "Task" [ 1127.324982] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.336227] env[63531]: DEBUG oslo_vmware.api [None req-26695b07-87c6-42d0-9fcc-9d1cd72db1af tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118698, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.397870] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ec05784d-d9c2-48ec-94cb-d6c915d23a53 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "5ad9fc53-104b-4a90-801a-bd1e8f12c1d7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1127.398187] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ec05784d-d9c2-48ec-94cb-d6c915d23a53 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "5ad9fc53-104b-4a90-801a-bd1e8f12c1d7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1127.398465] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ec05784d-d9c2-48ec-94cb-d6c915d23a53 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "5ad9fc53-104b-4a90-801a-bd1e8f12c1d7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1127.398706] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ec05784d-d9c2-48ec-94cb-d6c915d23a53 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "5ad9fc53-104b-4a90-801a-bd1e8f12c1d7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1127.398920] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ec05784d-d9c2-48ec-94cb-d6c915d23a53 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "5ad9fc53-104b-4a90-801a-bd1e8f12c1d7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1127.401191] env[63531]: INFO nova.compute.manager [None req-ec05784d-d9c2-48ec-94cb-d6c915d23a53 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] Terminating instance [ 1127.403110] env[63531]: DEBUG nova.compute.manager [None req-ec05784d-d9c2-48ec-94cb-d6c915d23a53 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1127.403492] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-ec05784d-d9c2-48ec-94cb-d6c915d23a53 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1127.404023] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4c090a2-38cf-4d6a-bc76-ed229204e08c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.412162] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec05784d-d9c2-48ec-94cb-d6c915d23a53 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1127.412416] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-314de965-fda3-434f-8964-80da70f6ee22 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.420576] env[63531]: DEBUG oslo_vmware.api [None req-ec05784d-d9c2-48ec-94cb-d6c915d23a53 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 1127.420576] env[63531]: value = "task-1118699" [ 1127.420576] env[63531]: _type = "Task" [ 1127.420576] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.429728] env[63531]: DEBUG oslo_vmware.api [None req-ec05784d-d9c2-48ec-94cb-d6c915d23a53 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118699, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.550187] env[63531]: DEBUG oslo_vmware.api [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]526a6a0d-cdf8-35e7-4e30-e4e06f418b69, 'name': SearchDatastore_Task, 'duration_secs': 0.009395} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.551512] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0c460ddc-6a91-46a3-892f-e48c733ee1e6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.557974] env[63531]: DEBUG oslo_vmware.api [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 1127.557974] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]5261bfc2-1f50-7bbb-6430-ce412b13e515" [ 1127.557974] env[63531]: _type = "Task" [ 1127.557974] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.566174] env[63531]: DEBUG oslo_vmware.api [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5261bfc2-1f50-7bbb-6430-ce412b13e515, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.652524] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 3bf08a09-b6e6-4a58-8d35-7061a8e59afd] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1127.668370] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c64a6390-4c15-48fc-b7ac-a0d9a2a8ddb5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquiring lock "8dec0cac-defc-42f2-ab56-6b3ae60ad858" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1127.668783] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c64a6390-4c15-48fc-b7ac-a0d9a2a8ddb5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "8dec0cac-defc-42f2-ab56-6b3ae60ad858" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1127.669048] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c64a6390-4c15-48fc-b7ac-a0d9a2a8ddb5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquiring lock "8dec0cac-defc-42f2-ab56-6b3ae60ad858-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1127.669258] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c64a6390-4c15-48fc-b7ac-a0d9a2a8ddb5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "8dec0cac-defc-42f2-ab56-6b3ae60ad858-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1127.669486] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c64a6390-4c15-48fc-b7ac-a0d9a2a8ddb5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "8dec0cac-defc-42f2-ab56-6b3ae60ad858-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1127.671724] env[63531]: INFO nova.compute.manager [None req-c64a6390-4c15-48fc-b7ac-a0d9a2a8ddb5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Terminating instance [ 1127.673806] env[63531]: DEBUG nova.compute.manager [None req-c64a6390-4c15-48fc-b7ac-a0d9a2a8ddb5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1127.674016] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c64a6390-4c15-48fc-b7ac-a0d9a2a8ddb5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1127.674828] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b08b5a4-d155-4e5f-9031-642549b1251d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.682804] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-c64a6390-4c15-48fc-b7ac-a0d9a2a8ddb5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1127.683044] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e80b65f6-759b-457f-8a7c-8e311589c1fe {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.690341] env[63531]: DEBUG oslo_vmware.api [None req-c64a6390-4c15-48fc-b7ac-a0d9a2a8ddb5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 1127.690341] env[63531]: value = "task-1118700" [ 1127.690341] env[63531]: _type = "Task" [ 1127.690341] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.700801] env[63531]: DEBUG oslo_vmware.api [None req-c64a6390-4c15-48fc-b7ac-a0d9a2a8ddb5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118700, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.757053] env[63531]: DEBUG oslo_concurrency.lockutils [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Releasing lock "refresh_cache-2d51997d-f88c-468c-9a56-0331a06ca56c" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1127.757400] env[63531]: DEBUG nova.compute.manager [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Instance network_info: |[{"id": "8c2307f2-c72e-47bb-a1f7-1dafecb0d2a4", "address": "fa:16:3e:54:75:66", "network": {"id": "0667c176-cfdf-4e43-b152-e00bdc224448", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-243715985-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "308eabafea9c4188a58a0f1c22074d2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8f580e6-1d86-41ee-9ebe-c531cb9299c6", "external-id": "nsx-vlan-transportzone-150", "segmentation_id": 150, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c2307f2-c7", "ovs_interfaceid": "8c2307f2-c72e-47bb-a1f7-1dafecb0d2a4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1127.757719] env[63531]: DEBUG oslo_concurrency.lockutils [req-ae346bcf-d1a4-4755-89b1-48f3545c8e02 req-9f2f97cc-5b51-4939-b4f1-51203fdddfde service nova] Acquired lock "refresh_cache-2d51997d-f88c-468c-9a56-0331a06ca56c" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1127.757905] env[63531]: DEBUG nova.network.neutron [req-ae346bcf-d1a4-4755-89b1-48f3545c8e02 req-9f2f97cc-5b51-4939-b4f1-51203fdddfde service nova] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Refreshing network info cache for port 8c2307f2-c72e-47bb-a1f7-1dafecb0d2a4 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1127.759955] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:54:75:66', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c8f580e6-1d86-41ee-9ebe-c531cb9299c6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8c2307f2-c72e-47bb-a1f7-1dafecb0d2a4', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1127.767154] env[63531]: DEBUG oslo.service.loopingcall [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1127.770293] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1127.770836] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-74e3e848-bd95-4da9-921a-b373851c47e6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.792651] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1127.792651] env[63531]: value = "task-1118701" [ 1127.792651] env[63531]: _type = "Task" [ 1127.792651] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.803198] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118701, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.832711] env[63531]: DEBUG nova.compute.manager [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Stashing vm_state: active {{(pid=63531) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1127.840777] env[63531]: DEBUG oslo_vmware.api [None req-26695b07-87c6-42d0-9fcc-9d1cd72db1af tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118698, 'name': PowerOffVM_Task, 'duration_secs': 0.204599} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.840777] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-26695b07-87c6-42d0-9fcc-9d1cd72db1af tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1127.840777] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-26695b07-87c6-42d0-9fcc-9d1cd72db1af tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1127.842035] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6aeca7fb-458d-4d18-a2aa-76bf4c23c57d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.933335] env[63531]: DEBUG oslo_vmware.api [None req-ec05784d-d9c2-48ec-94cb-d6c915d23a53 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118699, 'name': PowerOffVM_Task, 'duration_secs': 0.189924} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.937164] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec05784d-d9c2-48ec-94cb-d6c915d23a53 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1127.937330] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-ec05784d-d9c2-48ec-94cb-d6c915d23a53 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1127.937566] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bb3e55ee-b3fa-4cd7-9d97-117d257ad652 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.944038] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-26695b07-87c6-42d0-9fcc-9d1cd72db1af tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1127.944304] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-26695b07-87c6-42d0-9fcc-9d1cd72db1af tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1127.944457] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-26695b07-87c6-42d0-9fcc-9d1cd72db1af tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Deleting the datastore file [datastore1] 89f19669-53a9-4425-a3c6-d5f8873b707e {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1127.945315] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bc4013b5-484c-4a6b-85c0-e9e9f636f1a1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.954039] env[63531]: DEBUG oslo_vmware.api [None req-26695b07-87c6-42d0-9fcc-9d1cd72db1af tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for the task: (returnval){ [ 1127.954039] env[63531]: value = "task-1118704" [ 1127.954039] env[63531]: _type = "Task" [ 1127.954039] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.963633] env[63531]: DEBUG oslo_vmware.api [None req-26695b07-87c6-42d0-9fcc-9d1cd72db1af tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118704, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.016286] env[63531]: DEBUG nova.network.neutron [req-ae346bcf-d1a4-4755-89b1-48f3545c8e02 req-9f2f97cc-5b51-4939-b4f1-51203fdddfde service nova] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Updated VIF entry in instance network info cache for port 8c2307f2-c72e-47bb-a1f7-1dafecb0d2a4. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1128.016748] env[63531]: DEBUG nova.network.neutron [req-ae346bcf-d1a4-4755-89b1-48f3545c8e02 req-9f2f97cc-5b51-4939-b4f1-51203fdddfde service nova] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Updating instance_info_cache with network_info: [{"id": "8c2307f2-c72e-47bb-a1f7-1dafecb0d2a4", "address": "fa:16:3e:54:75:66", "network": {"id": "0667c176-cfdf-4e43-b152-e00bdc224448", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-243715985-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "308eabafea9c4188a58a0f1c22074d2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8f580e6-1d86-41ee-9ebe-c531cb9299c6", "external-id": "nsx-vlan-transportzone-150", "segmentation_id": 150, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c2307f2-c7", "ovs_interfaceid": "8c2307f2-c72e-47bb-a1f7-1dafecb0d2a4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1128.020515] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-ec05784d-d9c2-48ec-94cb-d6c915d23a53 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1128.020661] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-ec05784d-d9c2-48ec-94cb-d6c915d23a53 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] Deleting contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1128.021086] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec05784d-d9c2-48ec-94cb-d6c915d23a53 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Deleting the datastore file [datastore2] 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1128.021233] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dee7821d-0d11-4021-a249-9499e6c9b8bc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.030037] env[63531]: DEBUG oslo_vmware.api [None req-ec05784d-d9c2-48ec-94cb-d6c915d23a53 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for the task: (returnval){ [ 1128.030037] env[63531]: value = "task-1118705" [ 1128.030037] env[63531]: _type = "Task" [ 1128.030037] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.041300] env[63531]: DEBUG oslo_vmware.api [None req-ec05784d-d9c2-48ec-94cb-d6c915d23a53 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118705, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.069948] env[63531]: DEBUG oslo_vmware.api [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]5261bfc2-1f50-7bbb-6430-ce412b13e515, 'name': SearchDatastore_Task, 'duration_secs': 0.010532} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.070368] env[63531]: DEBUG oslo_concurrency.lockutils [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1128.070820] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97/3794f44c-8f6d-4bab-b9e0-9f6c819e1a97.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1128.071726] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6a82dcd3-dac4-4593-9986-697073a93201 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.081037] env[63531]: DEBUG oslo_vmware.api [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 1128.081037] env[63531]: value = "task-1118706" [ 1128.081037] env[63531]: _type = "Task" [ 1128.081037] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.091468] env[63531]: DEBUG oslo_vmware.api [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118706, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.156057] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: baf2db4e-2cdb-47f5-9a5c-97a233a459c9] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1128.200190] env[63531]: DEBUG oslo_vmware.api [None req-c64a6390-4c15-48fc-b7ac-a0d9a2a8ddb5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118700, 'name': PowerOffVM_Task, 'duration_secs': 0.209814} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.200446] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-c64a6390-4c15-48fc-b7ac-a0d9a2a8ddb5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1128.200625] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c64a6390-4c15-48fc-b7ac-a0d9a2a8ddb5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1128.200887] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5074f573-eb49-4795-9639-5d050c5aa941 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.273305] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c64a6390-4c15-48fc-b7ac-a0d9a2a8ddb5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1128.273516] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c64a6390-4c15-48fc-b7ac-a0d9a2a8ddb5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1128.273703] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-c64a6390-4c15-48fc-b7ac-a0d9a2a8ddb5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Deleting the datastore file [datastore1] 8dec0cac-defc-42f2-ab56-6b3ae60ad858 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1128.273973] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d19cebf2-93eb-4f1b-b26d-9f26926afd3a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.280527] env[63531]: DEBUG oslo_vmware.api [None req-c64a6390-4c15-48fc-b7ac-a0d9a2a8ddb5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for the task: (returnval){ [ 1128.280527] env[63531]: value = "task-1118708" [ 1128.280527] env[63531]: _type = "Task" [ 1128.280527] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.289221] env[63531]: DEBUG oslo_vmware.api [None req-c64a6390-4c15-48fc-b7ac-a0d9a2a8ddb5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118708, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.301531] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118701, 'name': CreateVM_Task, 'duration_secs': 0.36648} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.301687] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1128.302380] env[63531]: DEBUG oslo_concurrency.lockutils [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1128.302549] env[63531]: DEBUG oslo_concurrency.lockutils [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1128.302880] env[63531]: DEBUG oslo_concurrency.lockutils [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1128.303144] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-db8ba099-1760-48f5-9e09-e55f79f6f2d0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.307789] env[63531]: DEBUG oslo_vmware.api [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 1128.307789] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]525cfb63-4619-6d50-f94b-378c7a842d48" [ 1128.307789] env[63531]: _type = "Task" [ 1128.307789] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.317382] env[63531]: DEBUG oslo_vmware.api [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]525cfb63-4619-6d50-f94b-378c7a842d48, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.358427] env[63531]: DEBUG oslo_concurrency.lockutils [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1128.358727] env[63531]: DEBUG oslo_concurrency.lockutils [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1128.465838] env[63531]: DEBUG oslo_vmware.api [None req-26695b07-87c6-42d0-9fcc-9d1cd72db1af tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Task: {'id': task-1118704, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.168257} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.466140] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-26695b07-87c6-42d0-9fcc-9d1cd72db1af tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1128.466348] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-26695b07-87c6-42d0-9fcc-9d1cd72db1af tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1128.466538] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-26695b07-87c6-42d0-9fcc-9d1cd72db1af tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1128.466722] env[63531]: INFO nova.compute.manager [None req-26695b07-87c6-42d0-9fcc-9d1cd72db1af tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Took 1.17 seconds to destroy the instance on the hypervisor. [ 1128.466976] env[63531]: DEBUG oslo.service.loopingcall [None req-26695b07-87c6-42d0-9fcc-9d1cd72db1af tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1128.467217] env[63531]: DEBUG nova.compute.manager [-] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1128.467312] env[63531]: DEBUG nova.network.neutron [-] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1128.520017] env[63531]: DEBUG oslo_concurrency.lockutils [req-ae346bcf-d1a4-4755-89b1-48f3545c8e02 req-9f2f97cc-5b51-4939-b4f1-51203fdddfde service nova] Releasing lock "refresh_cache-2d51997d-f88c-468c-9a56-0331a06ca56c" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1128.543356] env[63531]: DEBUG oslo_vmware.api [None req-ec05784d-d9c2-48ec-94cb-d6c915d23a53 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Task: {'id': task-1118705, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.223345} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.545939] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec05784d-d9c2-48ec-94cb-d6c915d23a53 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1128.546157] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-ec05784d-d9c2-48ec-94cb-d6c915d23a53 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] Deleted contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1128.546396] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-ec05784d-d9c2-48ec-94cb-d6c915d23a53 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1128.546595] env[63531]: INFO nova.compute.manager [None req-ec05784d-d9c2-48ec-94cb-d6c915d23a53 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1128.546851] env[63531]: DEBUG oslo.service.loopingcall [None req-ec05784d-d9c2-48ec-94cb-d6c915d23a53 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1128.547129] env[63531]: DEBUG nova.compute.manager [-] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1128.547231] env[63531]: DEBUG nova.network.neutron [-] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1128.594387] env[63531]: DEBUG oslo_vmware.api [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118706, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.659887] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 0fb3b687-62cc-4296-b4ed-537aba18c880] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1128.790967] env[63531]: DEBUG oslo_vmware.api [None req-c64a6390-4c15-48fc-b7ac-a0d9a2a8ddb5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Task: {'id': task-1118708, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.445134} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.791303] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-c64a6390-4c15-48fc-b7ac-a0d9a2a8ddb5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1128.791517] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c64a6390-4c15-48fc-b7ac-a0d9a2a8ddb5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1128.791704] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c64a6390-4c15-48fc-b7ac-a0d9a2a8ddb5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1128.791884] env[63531]: INFO nova.compute.manager [None req-c64a6390-4c15-48fc-b7ac-a0d9a2a8ddb5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1128.792212] env[63531]: DEBUG oslo.service.loopingcall [None req-c64a6390-4c15-48fc-b7ac-a0d9a2a8ddb5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1128.792431] env[63531]: DEBUG nova.compute.manager [-] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1128.792554] env[63531]: DEBUG nova.network.neutron [-] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1128.819622] env[63531]: DEBUG oslo_vmware.api [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]525cfb63-4619-6d50-f94b-378c7a842d48, 'name': SearchDatastore_Task, 'duration_secs': 0.010815} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.819966] env[63531]: DEBUG oslo_concurrency.lockutils [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1128.820314] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1128.820609] env[63531]: DEBUG oslo_concurrency.lockutils [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1128.820762] env[63531]: DEBUG oslo_concurrency.lockutils [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1128.820944] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1128.821311] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f402bf46-3694-49de-8bb5-e903082bfe6d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.836690] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1128.836971] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1128.837776] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a7821528-94aa-4624-b164-357b608bf9e9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.843811] env[63531]: DEBUG oslo_vmware.api [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 1128.843811] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]525768c6-cdf1-84cf-545c-74ff0be318fa" [ 1128.843811] env[63531]: _type = "Task" [ 1128.843811] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.852816] env[63531]: DEBUG oslo_vmware.api [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]525768c6-cdf1-84cf-545c-74ff0be318fa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.865354] env[63531]: INFO nova.compute.claims [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1129.021159] env[63531]: DEBUG nova.compute.manager [req-fb5f6afb-498b-4f72-bc51-ce6275e97ca5 req-b915eee1-8d06-41bb-be77-adf3951b398c service nova] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] Received event network-vif-deleted-161e87a1-1133-476f-9598-44c29595c36a {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1129.021159] env[63531]: INFO nova.compute.manager [req-fb5f6afb-498b-4f72-bc51-ce6275e97ca5 req-b915eee1-8d06-41bb-be77-adf3951b398c service nova] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] Neutron deleted interface 161e87a1-1133-476f-9598-44c29595c36a; detaching it from the instance and deleting it from the info cache [ 1129.021159] env[63531]: DEBUG nova.network.neutron [req-fb5f6afb-498b-4f72-bc51-ce6275e97ca5 req-b915eee1-8d06-41bb-be77-adf3951b398c service nova] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1129.080106] env[63531]: DEBUG nova.compute.manager [req-d51371a5-2143-4782-8b78-05ed4d43cbb8 req-9c2e3952-3df6-451d-b5a6-58935205bae1 service nova] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Received event network-vif-deleted-76060df9-1ff6-4c64-86ef-a8e0d4081cfd {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1129.080106] env[63531]: INFO nova.compute.manager [req-d51371a5-2143-4782-8b78-05ed4d43cbb8 req-9c2e3952-3df6-451d-b5a6-58935205bae1 service nova] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Neutron deleted interface 76060df9-1ff6-4c64-86ef-a8e0d4081cfd; detaching it from the instance and deleting it from the info cache [ 1129.080106] env[63531]: DEBUG nova.network.neutron [req-d51371a5-2143-4782-8b78-05ed4d43cbb8 req-9c2e3952-3df6-451d-b5a6-58935205bae1 service nova] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1129.093867] env[63531]: DEBUG oslo_vmware.api [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118706, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.742348} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.093867] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97/3794f44c-8f6d-4bab-b9e0-9f6c819e1a97.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1129.093867] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1129.094136] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-65059af6-e1bd-4b99-a408-29c56feb5474 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.102963] env[63531]: DEBUG oslo_vmware.api [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 1129.102963] env[63531]: value = "task-1118709" [ 1129.102963] env[63531]: _type = "Task" [ 1129.102963] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.114575] env[63531]: DEBUG oslo_vmware.api [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118709, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.166266] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 6a9800ff-74f7-4e69-bf39-ac5e9bb52bbe] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1129.354594] env[63531]: DEBUG oslo_vmware.api [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]525768c6-cdf1-84cf-545c-74ff0be318fa, 'name': SearchDatastore_Task, 'duration_secs': 0.02833} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.357178] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-659d30e0-ec17-4db1-9918-2d953a412cb1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.361129] env[63531]: DEBUG oslo_vmware.api [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 1129.361129] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52dd653d-c0e3-0039-c0c0-3b3b8a59b612" [ 1129.361129] env[63531]: _type = "Task" [ 1129.361129] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.371656] env[63531]: DEBUG oslo_vmware.api [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52dd653d-c0e3-0039-c0c0-3b3b8a59b612, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.373256] env[63531]: INFO nova.compute.resource_tracker [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Updating resource usage from migration b80d81d1-28a5-4648-9e17-d1ccbade2da2 [ 1129.496463] env[63531]: DEBUG nova.network.neutron [-] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1129.509894] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-068efdd9-21d3-4e6a-aa48-0b898e1aeb00 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.517931] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f6c1010-b2be-4409-ba96-b4f6e8bbedf8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.548486] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6e16b4e4-a6ad-4815-b01f-153ed9b70cf6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.550700] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74addc0a-2e60-45f5-babc-a7346111f288 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.553558] env[63531]: DEBUG nova.network.neutron [-] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1129.559414] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77086287-0db5-484e-9393-719bd850bc1e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.566099] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8efb7456-f55e-40d5-9628-47dd10a51880 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.588278] env[63531]: DEBUG nova.compute.provider_tree [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Updating inventory in ProviderTree for provider 9996a314-a62e-4282-8252-f387e7e3fabb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1129.596242] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7222b586-31af-418a-ad68-f8a565e5bbf7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.598642] env[63531]: DEBUG nova.compute.manager [req-fb5f6afb-498b-4f72-bc51-ce6275e97ca5 req-b915eee1-8d06-41bb-be77-adf3951b398c service nova] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] Detach interface failed, port_id=161e87a1-1133-476f-9598-44c29595c36a, reason: Instance 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7 could not be found. {{(pid=63531) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1129.611172] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f34abe0-a708-4c26-8009-c88a2ef2167d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.622189] env[63531]: ERROR nova.scheduler.client.report [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [req-2df89602-05ab-4bdd-946e-eba764fb3f63] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 9996a314-a62e-4282-8252-f387e7e3fabb. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-2df89602-05ab-4bdd-946e-eba764fb3f63"}]} [ 1129.627257] env[63531]: DEBUG oslo_vmware.api [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118709, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.146334} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.628673] env[63531]: DEBUG nova.network.neutron [-] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1129.629398] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1129.630372] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0059607-7b82-419c-9671-15141bac8a57 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.655993] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Reconfiguring VM instance instance-00000072 to attach disk [datastore2] 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97/3794f44c-8f6d-4bab-b9e0-9f6c819e1a97.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1129.663238] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eb868d36-882b-42f3-bc35-608ab2a783a5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.677674] env[63531]: DEBUG nova.compute.manager [req-d51371a5-2143-4782-8b78-05ed4d43cbb8 req-9c2e3952-3df6-451d-b5a6-58935205bae1 service nova] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Detach interface failed, port_id=76060df9-1ff6-4c64-86ef-a8e0d4081cfd, reason: Instance 8dec0cac-defc-42f2-ab56-6b3ae60ad858 could not be found. {{(pid=63531) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1129.678559] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 911f23f7-e320-48f3-87a9-0239b013feff] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1129.690021] env[63531]: DEBUG oslo_vmware.api [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 1129.690021] env[63531]: value = "task-1118710" [ 1129.690021] env[63531]: _type = "Task" [ 1129.690021] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.696131] env[63531]: DEBUG oslo_vmware.api [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118710, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.733183] env[63531]: DEBUG nova.scheduler.client.report [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Refreshing inventories for resource provider 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1129.746822] env[63531]: DEBUG nova.scheduler.client.report [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Updating ProviderTree inventory for provider 9996a314-a62e-4282-8252-f387e7e3fabb from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1129.747077] env[63531]: DEBUG nova.compute.provider_tree [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Updating inventory in ProviderTree for provider 9996a314-a62e-4282-8252-f387e7e3fabb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 169, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1129.759091] env[63531]: DEBUG nova.scheduler.client.report [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Refreshing aggregate associations for resource provider 9996a314-a62e-4282-8252-f387e7e3fabb, aggregates: None {{(pid=63531) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1129.777880] env[63531]: DEBUG nova.scheduler.client.report [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Refreshing trait associations for resource provider 9996a314-a62e-4282-8252-f387e7e3fabb, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE {{(pid=63531) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1129.873858] env[63531]: DEBUG oslo_vmware.api [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52dd653d-c0e3-0039-c0c0-3b3b8a59b612, 'name': SearchDatastore_Task, 'duration_secs': 0.08516} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.874510] env[63531]: DEBUG oslo_concurrency.lockutils [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1129.874937] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] 2d51997d-f88c-468c-9a56-0331a06ca56c/2d51997d-f88c-468c-9a56-0331a06ca56c.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1129.874937] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-517ce9c7-6a3d-4b6c-8449-dff6f76e8178 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.886349] env[63531]: DEBUG oslo_vmware.api [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 1129.886349] env[63531]: value = "task-1118711" [ 1129.886349] env[63531]: _type = "Task" [ 1129.886349] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.897498] env[63531]: DEBUG oslo_vmware.api [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118711, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.904118] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d3fd87c-ff08-4634-bab7-fef829b030c6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.912308] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7ad1fd8-c88c-4c0c-bb4e-bc5ffa676d38 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.962705] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9f18692-99a8-4f31-9f46-5610839837fb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.971661] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17ce1f22-1572-4098-8c89-6048021afd51 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.985501] env[63531]: DEBUG nova.compute.provider_tree [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Updating inventory in ProviderTree for provider 9996a314-a62e-4282-8252-f387e7e3fabb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1129.999376] env[63531]: INFO nova.compute.manager [-] [instance: 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7] Took 1.45 seconds to deallocate network for instance. [ 1130.056661] env[63531]: INFO nova.compute.manager [-] [instance: 8dec0cac-defc-42f2-ab56-6b3ae60ad858] Took 1.26 seconds to deallocate network for instance. [ 1130.130687] env[63531]: INFO nova.compute.manager [-] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Took 1.66 seconds to deallocate network for instance. [ 1130.182061] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 55111b87-1e9a-4877-8689-987faa72a54e] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1130.198648] env[63531]: DEBUG oslo_vmware.api [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118710, 'name': ReconfigVM_Task, 'duration_secs': 0.302505} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.198945] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Reconfigured VM instance instance-00000072 to attach disk [datastore2] 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97/3794f44c-8f6d-4bab-b9e0-9f6c819e1a97.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1130.199556] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5c970463-194e-4cf8-a9d9-e2f7c3314bdb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.207056] env[63531]: DEBUG oslo_vmware.api [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 1130.207056] env[63531]: value = "task-1118712" [ 1130.207056] env[63531]: _type = "Task" [ 1130.207056] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.215478] env[63531]: DEBUG oslo_vmware.api [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118712, 'name': Rename_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.397590] env[63531]: DEBUG oslo_vmware.api [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118711, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.505249] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ec05784d-d9c2-48ec-94cb-d6c915d23a53 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1130.519960] env[63531]: DEBUG nova.scheduler.client.report [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Updated inventory for provider 9996a314-a62e-4282-8252-f387e7e3fabb with generation 161 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1130.520311] env[63531]: DEBUG nova.compute.provider_tree [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Updating resource provider 9996a314-a62e-4282-8252-f387e7e3fabb generation from 161 to 162 during operation: update_inventory {{(pid=63531) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1130.520508] env[63531]: DEBUG nova.compute.provider_tree [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Updating inventory in ProviderTree for provider 9996a314-a62e-4282-8252-f387e7e3fabb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1130.563840] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c64a6390-4c15-48fc-b7ac-a0d9a2a8ddb5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1130.637690] env[63531]: DEBUG oslo_concurrency.lockutils [None req-26695b07-87c6-42d0-9fcc-9d1cd72db1af tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1130.685888] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 27c23b7c-a68f-43c9-a125-bd657feb3c5b] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1130.718629] env[63531]: DEBUG oslo_vmware.api [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118712, 'name': Rename_Task, 'duration_secs': 0.235566} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.718915] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1130.719170] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b5e033cb-18e9-4010-bc73-6ec6a238345e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.725776] env[63531]: DEBUG oslo_vmware.api [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 1130.725776] env[63531]: value = "task-1118713" [ 1130.725776] env[63531]: _type = "Task" [ 1130.725776] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.734760] env[63531]: DEBUG oslo_vmware.api [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118713, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.898040] env[63531]: DEBUG oslo_vmware.api [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118711, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.645389} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.898168] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore2] 2d51997d-f88c-468c-9a56-0331a06ca56c/2d51997d-f88c-468c-9a56-0331a06ca56c.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1130.898529] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1130.898916] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-70dd78f7-4e7b-4393-afb3-b4384dd0a651 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.906070] env[63531]: DEBUG oslo_vmware.api [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 1130.906070] env[63531]: value = "task-1118714" [ 1130.906070] env[63531]: _type = "Task" [ 1130.906070] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.915250] env[63531]: DEBUG oslo_vmware.api [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118714, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.026038] env[63531]: DEBUG oslo_concurrency.lockutils [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.667s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1131.026293] env[63531]: INFO nova.compute.manager [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Migrating [ 1131.032918] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ec05784d-d9c2-48ec-94cb-d6c915d23a53 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.528s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1131.033177] env[63531]: DEBUG nova.objects.instance [None req-ec05784d-d9c2-48ec-94cb-d6c915d23a53 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lazy-loading 'resources' on Instance uuid 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1131.047187] env[63531]: DEBUG nova.compute.manager [req-e00494f5-236a-4a07-8d2d-65e50d754ad4 req-69de62f2-386c-4a38-b0b0-fc00f3862fb3 service nova] [instance: 89f19669-53a9-4425-a3c6-d5f8873b707e] Received event network-vif-deleted-bb17faa5-c291-4bfa-b2fa-c435a815dc3d {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1131.189171] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 3fb0499d-8074-4e63-be9f-380730416cc4] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1131.236739] env[63531]: DEBUG oslo_vmware.api [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118713, 'name': PowerOnVM_Task, 'duration_secs': 0.43708} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.237037] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1131.237254] env[63531]: INFO nova.compute.manager [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Took 7.84 seconds to spawn the instance on the hypervisor. [ 1131.237443] env[63531]: DEBUG nova.compute.manager [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1131.238249] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6b0cdb0-aa9a-4e5a-8549-8a772cbdd243 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.416739] env[63531]: DEBUG oslo_vmware.api [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118714, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071947} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.416967] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1131.417769] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d428288-afe7-45dd-9fc8-414b6bd8f738 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.439529] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Reconfiguring VM instance instance-00000073 to attach disk [datastore2] 2d51997d-f88c-468c-9a56-0331a06ca56c/2d51997d-f88c-468c-9a56-0331a06ca56c.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1131.439795] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3506949c-7381-47e0-b5c9-7f668d3dd4a7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.458986] env[63531]: DEBUG oslo_vmware.api [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 1131.458986] env[63531]: value = "task-1118715" [ 1131.458986] env[63531]: _type = "Task" [ 1131.458986] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.473384] env[63531]: DEBUG oslo_vmware.api [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118715, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.544835] env[63531]: DEBUG oslo_concurrency.lockutils [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "refresh_cache-225e11e5-0558-4ad5-acd4-2a69888fe2d1" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1131.545035] env[63531]: DEBUG oslo_concurrency.lockutils [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquired lock "refresh_cache-225e11e5-0558-4ad5-acd4-2a69888fe2d1" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1131.545219] env[63531]: DEBUG nova.network.neutron [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1131.657788] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5ebcf51-91bc-44f5-bf62-73fe36060a98 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.668203] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7217c025-f1ef-4631-83d4-7c15acc6b862 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.698185] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: ccf00b09-29fe-4a76-a8af-97774acd77de] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1131.700773] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae0bbc82-f15a-459a-bb80-5b555ce5588b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.709758] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d962274-660b-4073-8585-52ed091787be {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.725704] env[63531]: DEBUG nova.compute.provider_tree [None req-ec05784d-d9c2-48ec-94cb-d6c915d23a53 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1131.756954] env[63531]: INFO nova.compute.manager [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Took 17.20 seconds to build instance. [ 1131.969549] env[63531]: DEBUG oslo_vmware.api [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118715, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.204951] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: e065f052-ba3e-4783-8953-5dc200d1f3e9] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1132.228919] env[63531]: DEBUG nova.scheduler.client.report [None req-ec05784d-d9c2-48ec-94cb-d6c915d23a53 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1132.258655] env[63531]: DEBUG oslo_concurrency.lockutils [None req-97423453-c8f7-474e-86b9-e68498cb671d tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "3794f44c-8f6d-4bab-b9e0-9f6c819e1a97" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.709s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1132.366536] env[63531]: DEBUG nova.network.neutron [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Updating instance_info_cache with network_info: [{"id": "8b6b5f30-0bda-4851-9288-3c3a856e8a6c", "address": "fa:16:3e:8d:e8:d4", "network": {"id": "ea2d47d3-a502-4afd-9bc5-ee4a37017901", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1361282648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94a95667c00947ea9183a6307c569c90", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6c45fd-e930-495a-9cb7-df84eda443b1", "external-id": "nsx-vlan-transportzone-407", "segmentation_id": 407, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8b6b5f30-0b", "ovs_interfaceid": "8b6b5f30-0bda-4851-9288-3c3a856e8a6c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1132.469822] env[63531]: DEBUG oslo_vmware.api [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118715, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.708048] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 4b5e5821-9def-4b53-be19-355d9e5f81ec] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1132.735054] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ec05784d-d9c2-48ec-94cb-d6c915d23a53 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.702s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1132.737253] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c64a6390-4c15-48fc-b7ac-a0d9a2a8ddb5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.173s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1132.737494] env[63531]: DEBUG nova.objects.instance [None req-c64a6390-4c15-48fc-b7ac-a0d9a2a8ddb5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lazy-loading 'resources' on Instance uuid 8dec0cac-defc-42f2-ab56-6b3ae60ad858 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1132.754093] env[63531]: INFO nova.scheduler.client.report [None req-ec05784d-d9c2-48ec-94cb-d6c915d23a53 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Deleted allocations for instance 5ad9fc53-104b-4a90-801a-bd1e8f12c1d7 [ 1132.869585] env[63531]: DEBUG oslo_concurrency.lockutils [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Releasing lock "refresh_cache-225e11e5-0558-4ad5-acd4-2a69888fe2d1" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1132.970736] env[63531]: DEBUG oslo_vmware.api [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118715, 'name': ReconfigVM_Task, 'duration_secs': 1.034127} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.971024] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Reconfigured VM instance instance-00000073 to attach disk [datastore2] 2d51997d-f88c-468c-9a56-0331a06ca56c/2d51997d-f88c-468c-9a56-0331a06ca56c.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1132.971706] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-01d39216-802e-46dc-b625-9ef471f655f0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.978466] env[63531]: DEBUG oslo_vmware.api [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 1132.978466] env[63531]: value = "task-1118716" [ 1132.978466] env[63531]: _type = "Task" [ 1132.978466] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.985866] env[63531]: DEBUG oslo_vmware.api [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118716, 'name': Rename_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.086351] env[63531]: DEBUG nova.compute.manager [req-cab7d7ce-207f-4266-b3d0-4c5b70e9edc4 req-3a970f51-8c4f-424e-bb47-4447db0e5ad5 service nova] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Received event network-changed-9a3bf9ef-a105-4821-81f8-4891f7c3ae0a {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1133.086351] env[63531]: DEBUG nova.compute.manager [req-cab7d7ce-207f-4266-b3d0-4c5b70e9edc4 req-3a970f51-8c4f-424e-bb47-4447db0e5ad5 service nova] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Refreshing instance network info cache due to event network-changed-9a3bf9ef-a105-4821-81f8-4891f7c3ae0a. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1133.086522] env[63531]: DEBUG oslo_concurrency.lockutils [req-cab7d7ce-207f-4266-b3d0-4c5b70e9edc4 req-3a970f51-8c4f-424e-bb47-4447db0e5ad5 service nova] Acquiring lock "refresh_cache-3794f44c-8f6d-4bab-b9e0-9f6c819e1a97" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1133.086587] env[63531]: DEBUG oslo_concurrency.lockutils [req-cab7d7ce-207f-4266-b3d0-4c5b70e9edc4 req-3a970f51-8c4f-424e-bb47-4447db0e5ad5 service nova] Acquired lock "refresh_cache-3794f44c-8f6d-4bab-b9e0-9f6c819e1a97" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1133.086757] env[63531]: DEBUG nova.network.neutron [req-cab7d7ce-207f-4266-b3d0-4c5b70e9edc4 req-3a970f51-8c4f-424e-bb47-4447db0e5ad5 service nova] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Refreshing network info cache for port 9a3bf9ef-a105-4821-81f8-4891f7c3ae0a {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1133.211696] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 06ec91b2-14cb-4db8-8fab-e6c4183b1116] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1133.261933] env[63531]: DEBUG oslo_concurrency.lockutils [None req-ec05784d-d9c2-48ec-94cb-d6c915d23a53 tempest-ServersTestJSON-314157237 tempest-ServersTestJSON-314157237-project-member] Lock "5ad9fc53-104b-4a90-801a-bd1e8f12c1d7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.864s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1133.346208] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57b2821b-6001-422d-9661-fc1268a1a3fa {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.354617] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2373348-add4-4acb-93f1-05e547ed4201 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.389296] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d0b9091-b485-441a-87e6-4580a26f9d5e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.397084] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee3e6f6d-d1f8-4320-b3f0-a657ffa681b0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.410879] env[63531]: DEBUG nova.compute.provider_tree [None req-c64a6390-4c15-48fc-b7ac-a0d9a2a8ddb5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1133.487856] env[63531]: DEBUG oslo_vmware.api [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118716, 'name': Rename_Task, 'duration_secs': 0.153022} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.488257] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1133.488418] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-66f3403f-edbc-4522-91c9-d43d5b6702cd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.495418] env[63531]: DEBUG oslo_vmware.api [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 1133.495418] env[63531]: value = "task-1118717" [ 1133.495418] env[63531]: _type = "Task" [ 1133.495418] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.502723] env[63531]: DEBUG oslo_vmware.api [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118717, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.716828] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 6118e242-c0db-4707-a5f9-3d12b823935f] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1133.914303] env[63531]: DEBUG nova.scheduler.client.report [None req-c64a6390-4c15-48fc-b7ac-a0d9a2a8ddb5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1134.012672] env[63531]: DEBUG oslo_vmware.api [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118717, 'name': PowerOnVM_Task} progress is 94%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.199782] env[63531]: DEBUG nova.network.neutron [req-cab7d7ce-207f-4266-b3d0-4c5b70e9edc4 req-3a970f51-8c4f-424e-bb47-4447db0e5ad5 service nova] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Updated VIF entry in instance network info cache for port 9a3bf9ef-a105-4821-81f8-4891f7c3ae0a. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1134.200182] env[63531]: DEBUG nova.network.neutron [req-cab7d7ce-207f-4266-b3d0-4c5b70e9edc4 req-3a970f51-8c4f-424e-bb47-4447db0e5ad5 service nova] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Updating instance_info_cache with network_info: [{"id": "9a3bf9ef-a105-4821-81f8-4891f7c3ae0a", "address": "fa:16:3e:dd:5e:bd", "network": {"id": "c4f4c994-36ab-42ab-94f8-78f29cbafa50", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-994507698-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "57e9bf4ee0f1463fa67b0f8deff30fdc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "767a3a48-41d4-4a0c-961d-0024837f63bd", "external-id": "nsx-vlan-transportzone-141", "segmentation_id": 141, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a3bf9ef-a1", "ovs_interfaceid": "9a3bf9ef-a105-4821-81f8-4891f7c3ae0a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1134.219371] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: dd7a850c-eeaa-49a6-a2b9-4d56b6748e5a] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1134.402023] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d683d5f5-b942-4860-9777-a609bc6a1d69 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.419665] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c64a6390-4c15-48fc-b7ac-a0d9a2a8ddb5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.682s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1134.422427] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Updating instance '225e11e5-0558-4ad5-acd4-2a69888fe2d1' progress to 0 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1134.426954] env[63531]: DEBUG oslo_concurrency.lockutils [None req-26695b07-87c6-42d0-9fcc-9d1cd72db1af tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.789s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1134.427317] env[63531]: DEBUG nova.objects.instance [None req-26695b07-87c6-42d0-9fcc-9d1cd72db1af tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lazy-loading 'resources' on Instance uuid 89f19669-53a9-4425-a3c6-d5f8873b707e {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1134.448621] env[63531]: INFO nova.scheduler.client.report [None req-c64a6390-4c15-48fc-b7ac-a0d9a2a8ddb5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Deleted allocations for instance 8dec0cac-defc-42f2-ab56-6b3ae60ad858 [ 1134.506269] env[63531]: DEBUG oslo_vmware.api [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118717, 'name': PowerOnVM_Task, 'duration_secs': 0.563207} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.506581] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1134.506843] env[63531]: INFO nova.compute.manager [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Took 8.79 seconds to spawn the instance on the hypervisor. [ 1134.506933] env[63531]: DEBUG nova.compute.manager [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1134.508062] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c14b72ff-adc5-4026-a284-c492f30705b3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.703049] env[63531]: DEBUG oslo_concurrency.lockutils [req-cab7d7ce-207f-4266-b3d0-4c5b70e9edc4 req-3a970f51-8c4f-424e-bb47-4447db0e5ad5 service nova] Releasing lock "refresh_cache-3794f44c-8f6d-4bab-b9e0-9f6c819e1a97" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1134.722810] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: a090135f-489d-457d-be36-ba6f61b71ab8] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1134.931840] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1134.934647] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9b9e24cf-ba44-4a51-b44f-5904b9049d82 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.944159] env[63531]: DEBUG oslo_vmware.api [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 1134.944159] env[63531]: value = "task-1118718" [ 1134.944159] env[63531]: _type = "Task" [ 1134.944159] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.956754] env[63531]: DEBUG oslo_vmware.api [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118718, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.961012] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c64a6390-4c15-48fc-b7ac-a0d9a2a8ddb5 tempest-ServerActionsTestOtherA-1126549824 tempest-ServerActionsTestOtherA-1126549824-project-member] Lock "8dec0cac-defc-42f2-ab56-6b3ae60ad858" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.292s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1135.031048] env[63531]: INFO nova.compute.manager [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Took 20.38 seconds to build instance. [ 1135.070960] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9dce500-1064-4209-bf77-88995aef464a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.079323] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a43dc03-f695-4fb1-80e5-be1cc87b815d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.117588] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-083cebf7-2b69-47ee-a591-a27c44d92d51 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.125010] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abcb3be9-0fae-496c-aa6b-45aa5e5b14cf {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.139779] env[63531]: DEBUG nova.compute.provider_tree [None req-26695b07-87c6-42d0-9fcc-9d1cd72db1af tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1135.226551] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 37658080-1d8a-46d5-ba6d-a5f0c0ce3fe9] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1135.459796] env[63531]: DEBUG oslo_vmware.api [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118718, 'name': PowerOffVM_Task, 'duration_secs': 0.285127} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.460048] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1135.460251] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Updating instance '225e11e5-0558-4ad5-acd4-2a69888fe2d1' progress to 17 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1135.531928] env[63531]: DEBUG oslo_concurrency.lockutils [None req-26a9dcca-cd43-4813-9a3d-2b47d3edfbb0 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "2d51997d-f88c-468c-9a56-0331a06ca56c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.894s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1135.643810] env[63531]: DEBUG nova.scheduler.client.report [None req-26695b07-87c6-42d0-9fcc-9d1cd72db1af tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1135.731407] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: e95d3273-8216-47cc-95b6-99301366a827] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1135.970589] env[63531]: DEBUG nova.virt.hardware [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1135.971244] env[63531]: DEBUG nova.virt.hardware [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1135.971244] env[63531]: DEBUG nova.virt.hardware [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1135.971553] env[63531]: DEBUG nova.virt.hardware [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1135.972598] env[63531]: DEBUG nova.virt.hardware [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1135.972598] env[63531]: DEBUG nova.virt.hardware [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1135.972598] env[63531]: DEBUG nova.virt.hardware [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1135.972598] env[63531]: DEBUG nova.virt.hardware [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1135.972598] env[63531]: DEBUG nova.virt.hardware [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1135.972830] env[63531]: DEBUG nova.virt.hardware [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1135.973094] env[63531]: DEBUG nova.virt.hardware [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1135.978654] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c1a8c03e-3069-416f-88fb-c432397c4b7c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.993669] env[63531]: DEBUG nova.compute.manager [req-fcee4453-5003-45db-9f65-7636d7f5fb93 req-c0f0c9f7-1b6f-478a-9597-4b5267d3fcf8 service nova] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Received event network-changed-8c2307f2-c72e-47bb-a1f7-1dafecb0d2a4 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1135.993887] env[63531]: DEBUG nova.compute.manager [req-fcee4453-5003-45db-9f65-7636d7f5fb93 req-c0f0c9f7-1b6f-478a-9597-4b5267d3fcf8 service nova] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Refreshing instance network info cache due to event network-changed-8c2307f2-c72e-47bb-a1f7-1dafecb0d2a4. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1135.994167] env[63531]: DEBUG oslo_concurrency.lockutils [req-fcee4453-5003-45db-9f65-7636d7f5fb93 req-c0f0c9f7-1b6f-478a-9597-4b5267d3fcf8 service nova] Acquiring lock "refresh_cache-2d51997d-f88c-468c-9a56-0331a06ca56c" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1135.994276] env[63531]: DEBUG oslo_concurrency.lockutils [req-fcee4453-5003-45db-9f65-7636d7f5fb93 req-c0f0c9f7-1b6f-478a-9597-4b5267d3fcf8 service nova] Acquired lock "refresh_cache-2d51997d-f88c-468c-9a56-0331a06ca56c" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1135.994406] env[63531]: DEBUG nova.network.neutron [req-fcee4453-5003-45db-9f65-7636d7f5fb93 req-c0f0c9f7-1b6f-478a-9597-4b5267d3fcf8 service nova] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Refreshing network info cache for port 8c2307f2-c72e-47bb-a1f7-1dafecb0d2a4 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1135.997780] env[63531]: DEBUG oslo_vmware.api [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 1135.997780] env[63531]: value = "task-1118719" [ 1135.997780] env[63531]: _type = "Task" [ 1135.997780] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.013922] env[63531]: DEBUG oslo_vmware.api [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118719, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.149197] env[63531]: DEBUG oslo_concurrency.lockutils [None req-26695b07-87c6-42d0-9fcc-9d1cd72db1af tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.722s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1136.175353] env[63531]: INFO nova.scheduler.client.report [None req-26695b07-87c6-42d0-9fcc-9d1cd72db1af tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Deleted allocations for instance 89f19669-53a9-4425-a3c6-d5f8873b707e [ 1136.239980] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 879924d4-b465-4102-a0e3-c7b2be7ef08b] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1136.511835] env[63531]: DEBUG oslo_vmware.api [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118719, 'name': ReconfigVM_Task, 'duration_secs': 0.238598} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.512366] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Updating instance '225e11e5-0558-4ad5-acd4-2a69888fe2d1' progress to 33 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1136.684692] env[63531]: DEBUG oslo_concurrency.lockutils [None req-26695b07-87c6-42d0-9fcc-9d1cd72db1af tempest-AttachVolumeNegativeTest-230851450 tempest-AttachVolumeNegativeTest-230851450-project-member] Lock "89f19669-53a9-4425-a3c6-d5f8873b707e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.389s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1136.743803] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 4b9ba2f0-d0c8-4d88-98f4-fe74f2d012df] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1136.891166] env[63531]: DEBUG nova.network.neutron [req-fcee4453-5003-45db-9f65-7636d7f5fb93 req-c0f0c9f7-1b6f-478a-9597-4b5267d3fcf8 service nova] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Updated VIF entry in instance network info cache for port 8c2307f2-c72e-47bb-a1f7-1dafecb0d2a4. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1136.891166] env[63531]: DEBUG nova.network.neutron [req-fcee4453-5003-45db-9f65-7636d7f5fb93 req-c0f0c9f7-1b6f-478a-9597-4b5267d3fcf8 service nova] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Updating instance_info_cache with network_info: [{"id": "8c2307f2-c72e-47bb-a1f7-1dafecb0d2a4", "address": "fa:16:3e:54:75:66", "network": {"id": "0667c176-cfdf-4e43-b152-e00bdc224448", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-243715985-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.217", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "308eabafea9c4188a58a0f1c22074d2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8f580e6-1d86-41ee-9ebe-c531cb9299c6", "external-id": "nsx-vlan-transportzone-150", "segmentation_id": 150, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c2307f2-c7", "ovs_interfaceid": "8c2307f2-c72e-47bb-a1f7-1dafecb0d2a4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1136.989322] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0043fe24-1de4-469e-b8fa-24bb04ed3c98 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "interface-ccc8f366-5849-45a6-9745-ff9b2be9c3bb-baf61e1c-97ff-411c-a0f7-415896653c60" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1136.989639] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0043fe24-1de4-469e-b8fa-24bb04ed3c98 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "interface-ccc8f366-5849-45a6-9745-ff9b2be9c3bb-baf61e1c-97ff-411c-a0f7-415896653c60" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1136.990040] env[63531]: DEBUG nova.objects.instance [None req-0043fe24-1de4-469e-b8fa-24bb04ed3c98 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lazy-loading 'flavor' on Instance uuid ccc8f366-5849-45a6-9745-ff9b2be9c3bb {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1137.018877] env[63531]: DEBUG nova.virt.hardware [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1137.019178] env[63531]: DEBUG nova.virt.hardware [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1137.019380] env[63531]: DEBUG nova.virt.hardware [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1137.019602] env[63531]: DEBUG nova.virt.hardware [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1137.019783] env[63531]: DEBUG nova.virt.hardware [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1137.019956] env[63531]: DEBUG nova.virt.hardware [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1137.020227] env[63531]: DEBUG nova.virt.hardware [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1137.020469] env[63531]: DEBUG nova.virt.hardware [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1137.020685] env[63531]: DEBUG nova.virt.hardware [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1137.023213] env[63531]: DEBUG nova.virt.hardware [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1137.023213] env[63531]: DEBUG nova.virt.hardware [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1137.026962] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Reconfiguring VM instance instance-00000071 to detach disk 2000 {{(pid=63531) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1137.028017] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d3c2848e-aead-48b5-87c3-36cd65158175 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.053242] env[63531]: DEBUG oslo_vmware.api [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 1137.053242] env[63531]: value = "task-1118720" [ 1137.053242] env[63531]: _type = "Task" [ 1137.053242] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.063454] env[63531]: DEBUG oslo_vmware.api [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118720, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.249387] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: db83adf4-6183-4330-b260-77d1f5daf899] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1137.393326] env[63531]: DEBUG oslo_concurrency.lockutils [req-fcee4453-5003-45db-9f65-7636d7f5fb93 req-c0f0c9f7-1b6f-478a-9597-4b5267d3fcf8 service nova] Releasing lock "refresh_cache-2d51997d-f88c-468c-9a56-0331a06ca56c" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1137.564313] env[63531]: DEBUG oslo_vmware.api [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118720, 'name': ReconfigVM_Task, 'duration_secs': 0.179546} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.564589] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Reconfigured VM instance instance-00000071 to detach disk 2000 {{(pid=63531) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1137.565389] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd1eb33e-8f54-4b91-9e83-949cf56489fb {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.589560] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Reconfiguring VM instance instance-00000071 to attach disk [datastore2] 225e11e5-0558-4ad5-acd4-2a69888fe2d1/225e11e5-0558-4ad5-acd4-2a69888fe2d1.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1137.589863] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-edd295e8-6564-4523-9d41-6ba0a448986b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.603278] env[63531]: DEBUG nova.objects.instance [None req-0043fe24-1de4-469e-b8fa-24bb04ed3c98 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lazy-loading 'pci_requests' on Instance uuid ccc8f366-5849-45a6-9745-ff9b2be9c3bb {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1137.610766] env[63531]: DEBUG oslo_vmware.api [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 1137.610766] env[63531]: value = "task-1118722" [ 1137.610766] env[63531]: _type = "Task" [ 1137.610766] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.619297] env[63531]: DEBUG oslo_vmware.api [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118722, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.752983] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 92a526e8-fddc-4ce0-8aab-dbdf4ae7ae1b] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1138.105814] env[63531]: DEBUG nova.objects.base [None req-0043fe24-1de4-469e-b8fa-24bb04ed3c98 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=63531) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1138.106068] env[63531]: DEBUG nova.network.neutron [None req-0043fe24-1de4-469e-b8fa-24bb04ed3c98 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1138.120657] env[63531]: DEBUG oslo_vmware.api [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118722, 'name': ReconfigVM_Task, 'duration_secs': 0.270757} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.120946] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Reconfigured VM instance instance-00000071 to attach disk [datastore2] 225e11e5-0558-4ad5-acd4-2a69888fe2d1/225e11e5-0558-4ad5-acd4-2a69888fe2d1.vmdk or device None with type thin {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1138.121241] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Updating instance '225e11e5-0558-4ad5-acd4-2a69888fe2d1' progress to 50 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1138.176024] env[63531]: DEBUG nova.policy [None req-0043fe24-1de4-469e-b8fa-24bb04ed3c98 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '36cb2f0de54f417384fa1fb2210de4e4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8ba8b284854242c392aec5326e996239', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 1138.256505] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 2b0e4285-310a-439c-bc50-000a7c5ef7f9] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1138.628688] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6b30e71-4670-4efe-8033-66c188540d52 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.650200] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72e08ae4-6cdc-48f0-a858-e31153528cd3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.669330] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Updating instance '225e11e5-0558-4ad5-acd4-2a69888fe2d1' progress to 67 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1138.760564] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 09feb12a-da9f-4bb7-959d-84c5c7c58e99] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1139.259674] env[63531]: DEBUG nova.network.neutron [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Port 8b6b5f30-0bda-4851-9288-3c3a856e8a6c binding to destination host cpu-1 is already ACTIVE {{(pid=63531) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1139.264700] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 3eeaa6f9-60d9-43fe-a438-fbf2ceececae] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1139.646676] env[63531]: DEBUG nova.compute.manager [req-7b0c6912-6a26-41c5-970b-083a1edf1909 req-4f1f60c4-253e-4e31-a7ea-40d9eeb340d2 service nova] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Received event network-vif-plugged-baf61e1c-97ff-411c-a0f7-415896653c60 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1139.646910] env[63531]: DEBUG oslo_concurrency.lockutils [req-7b0c6912-6a26-41c5-970b-083a1edf1909 req-4f1f60c4-253e-4e31-a7ea-40d9eeb340d2 service nova] Acquiring lock "ccc8f366-5849-45a6-9745-ff9b2be9c3bb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1139.647146] env[63531]: DEBUG oslo_concurrency.lockutils [req-7b0c6912-6a26-41c5-970b-083a1edf1909 req-4f1f60c4-253e-4e31-a7ea-40d9eeb340d2 service nova] Lock "ccc8f366-5849-45a6-9745-ff9b2be9c3bb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1139.647321] env[63531]: DEBUG oslo_concurrency.lockutils [req-7b0c6912-6a26-41c5-970b-083a1edf1909 req-4f1f60c4-253e-4e31-a7ea-40d9eeb340d2 service nova] Lock "ccc8f366-5849-45a6-9745-ff9b2be9c3bb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1139.647498] env[63531]: DEBUG nova.compute.manager [req-7b0c6912-6a26-41c5-970b-083a1edf1909 req-4f1f60c4-253e-4e31-a7ea-40d9eeb340d2 service nova] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] No waiting events found dispatching network-vif-plugged-baf61e1c-97ff-411c-a0f7-415896653c60 {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1139.647671] env[63531]: WARNING nova.compute.manager [req-7b0c6912-6a26-41c5-970b-083a1edf1909 req-4f1f60c4-253e-4e31-a7ea-40d9eeb340d2 service nova] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Received unexpected event network-vif-plugged-baf61e1c-97ff-411c-a0f7-415896653c60 for instance with vm_state active and task_state None. [ 1139.719379] env[63531]: DEBUG nova.network.neutron [None req-0043fe24-1de4-469e-b8fa-24bb04ed3c98 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Successfully updated port: baf61e1c-97ff-411c-a0f7-415896653c60 {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1139.771356] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 56d804bc-82a5-4ff7-a68f-c9a82e4e7d8d] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1140.223056] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0043fe24-1de4-469e-b8fa-24bb04ed3c98 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "refresh_cache-ccc8f366-5849-45a6-9745-ff9b2be9c3bb" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1140.223248] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0043fe24-1de4-469e-b8fa-24bb04ed3c98 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquired lock "refresh_cache-ccc8f366-5849-45a6-9745-ff9b2be9c3bb" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1140.223445] env[63531]: DEBUG nova.network.neutron [None req-0043fe24-1de4-469e-b8fa-24bb04ed3c98 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1140.279338] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 45b05c54-6865-47a1-ba93-90ad3e1ba07e] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1140.284179] env[63531]: DEBUG oslo_concurrency.lockutils [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "225e11e5-0558-4ad5-acd4-2a69888fe2d1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1140.284424] env[63531]: DEBUG oslo_concurrency.lockutils [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "225e11e5-0558-4ad5-acd4-2a69888fe2d1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1140.284611] env[63531]: DEBUG oslo_concurrency.lockutils [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "225e11e5-0558-4ad5-acd4-2a69888fe2d1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1140.759310] env[63531]: WARNING nova.network.neutron [None req-0043fe24-1de4-469e-b8fa-24bb04ed3c98 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] caf86428-59a5-4e50-9135-2ab8473af64d already exists in list: networks containing: ['caf86428-59a5-4e50-9135-2ab8473af64d']. ignoring it [ 1140.789022] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 43c29443-1210-4ee2-95c7-1257de308287] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1141.026891] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Acquiring lock "4de350e2-55cc-4795-85e6-e58b5b362459" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1141.027164] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Lock "4de350e2-55cc-4795-85e6-e58b5b362459" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1141.062496] env[63531]: DEBUG nova.network.neutron [None req-0043fe24-1de4-469e-b8fa-24bb04ed3c98 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Updating instance_info_cache with network_info: [{"id": "e9692fb2-f3a9-455c-8317-b533a6af48f7", "address": "fa:16:3e:79:4a:3b", "network": {"id": "caf86428-59a5-4e50-9135-2ab8473af64d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1673994696-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ba8b284854242c392aec5326e996239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape9692fb2-f3", "ovs_interfaceid": "e9692fb2-f3a9-455c-8317-b533a6af48f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "baf61e1c-97ff-411c-a0f7-415896653c60", "address": "fa:16:3e:88:db:88", "network": {"id": "caf86428-59a5-4e50-9135-2ab8473af64d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1673994696-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ba8b284854242c392aec5326e996239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbaf61e1c-97", "ovs_interfaceid": "baf61e1c-97ff-411c-a0f7-415896653c60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1141.291206] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: e0074e2e-e8bd-4bcb-9ad9-271a10e079e4] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1141.325729] env[63531]: DEBUG oslo_concurrency.lockutils [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "refresh_cache-225e11e5-0558-4ad5-acd4-2a69888fe2d1" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1141.325929] env[63531]: DEBUG oslo_concurrency.lockutils [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquired lock "refresh_cache-225e11e5-0558-4ad5-acd4-2a69888fe2d1" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1141.326126] env[63531]: DEBUG nova.network.neutron [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1141.530734] env[63531]: DEBUG nova.compute.manager [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] [instance: 4de350e2-55cc-4795-85e6-e58b5b362459] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1141.565279] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0043fe24-1de4-469e-b8fa-24bb04ed3c98 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Releasing lock "refresh_cache-ccc8f366-5849-45a6-9745-ff9b2be9c3bb" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1141.566230] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0043fe24-1de4-469e-b8fa-24bb04ed3c98 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "ccc8f366-5849-45a6-9745-ff9b2be9c3bb" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1141.566411] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0043fe24-1de4-469e-b8fa-24bb04ed3c98 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquired lock "ccc8f366-5849-45a6-9745-ff9b2be9c3bb" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1141.567393] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbc20aad-3fd7-489f-8fe8-11bd2b172052 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.587425] env[63531]: DEBUG nova.virt.hardware [None req-0043fe24-1de4-469e-b8fa-24bb04ed3c98 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1141.587688] env[63531]: DEBUG nova.virt.hardware [None req-0043fe24-1de4-469e-b8fa-24bb04ed3c98 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1141.587933] env[63531]: DEBUG nova.virt.hardware [None req-0043fe24-1de4-469e-b8fa-24bb04ed3c98 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1141.588043] env[63531]: DEBUG nova.virt.hardware [None req-0043fe24-1de4-469e-b8fa-24bb04ed3c98 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1141.588195] env[63531]: DEBUG nova.virt.hardware [None req-0043fe24-1de4-469e-b8fa-24bb04ed3c98 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1141.588345] env[63531]: DEBUG nova.virt.hardware [None req-0043fe24-1de4-469e-b8fa-24bb04ed3c98 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1141.588593] env[63531]: DEBUG nova.virt.hardware [None req-0043fe24-1de4-469e-b8fa-24bb04ed3c98 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1141.588774] env[63531]: DEBUG nova.virt.hardware [None req-0043fe24-1de4-469e-b8fa-24bb04ed3c98 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1141.589916] env[63531]: DEBUG nova.virt.hardware [None req-0043fe24-1de4-469e-b8fa-24bb04ed3c98 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1141.589916] env[63531]: DEBUG nova.virt.hardware [None req-0043fe24-1de4-469e-b8fa-24bb04ed3c98 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1141.589916] env[63531]: DEBUG nova.virt.hardware [None req-0043fe24-1de4-469e-b8fa-24bb04ed3c98 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1141.595569] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-0043fe24-1de4-469e-b8fa-24bb04ed3c98 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Reconfiguring VM to attach interface {{(pid=63531) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 1141.596242] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-52c6e79e-e067-4c9f-a95e-2d87ef75d6fe {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.614288] env[63531]: DEBUG oslo_vmware.api [None req-0043fe24-1de4-469e-b8fa-24bb04ed3c98 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for the task: (returnval){ [ 1141.614288] env[63531]: value = "task-1118724" [ 1141.614288] env[63531]: _type = "Task" [ 1141.614288] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1141.622144] env[63531]: DEBUG oslo_vmware.api [None req-0043fe24-1de4-469e-b8fa-24bb04ed3c98 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118724, 'name': ReconfigVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.690645] env[63531]: DEBUG nova.compute.manager [req-95199f1c-5707-4708-99ec-3b3df3f1afe6 req-8e6245b0-6fdc-46c3-a17c-d954f4bb994c service nova] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Received event network-changed-baf61e1c-97ff-411c-a0f7-415896653c60 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1141.690862] env[63531]: DEBUG nova.compute.manager [req-95199f1c-5707-4708-99ec-3b3df3f1afe6 req-8e6245b0-6fdc-46c3-a17c-d954f4bb994c service nova] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Refreshing instance network info cache due to event network-changed-baf61e1c-97ff-411c-a0f7-415896653c60. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1141.691109] env[63531]: DEBUG oslo_concurrency.lockutils [req-95199f1c-5707-4708-99ec-3b3df3f1afe6 req-8e6245b0-6fdc-46c3-a17c-d954f4bb994c service nova] Acquiring lock "refresh_cache-ccc8f366-5849-45a6-9745-ff9b2be9c3bb" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1141.691262] env[63531]: DEBUG oslo_concurrency.lockutils [req-95199f1c-5707-4708-99ec-3b3df3f1afe6 req-8e6245b0-6fdc-46c3-a17c-d954f4bb994c service nova] Acquired lock "refresh_cache-ccc8f366-5849-45a6-9745-ff9b2be9c3bb" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1141.691427] env[63531]: DEBUG nova.network.neutron [req-95199f1c-5707-4708-99ec-3b3df3f1afe6 req-8e6245b0-6fdc-46c3-a17c-d954f4bb994c service nova] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Refreshing network info cache for port baf61e1c-97ff-411c-a0f7-415896653c60 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1141.795037] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: f46d8a85-6ab7-436a-bc5e-40d0f80974e5] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1142.031480] env[63531]: DEBUG nova.network.neutron [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Updating instance_info_cache with network_info: [{"id": "8b6b5f30-0bda-4851-9288-3c3a856e8a6c", "address": "fa:16:3e:8d:e8:d4", "network": {"id": "ea2d47d3-a502-4afd-9bc5-ee4a37017901", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1361282648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94a95667c00947ea9183a6307c569c90", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6c45fd-e930-495a-9cb7-df84eda443b1", "external-id": "nsx-vlan-transportzone-407", "segmentation_id": 407, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8b6b5f30-0b", "ovs_interfaceid": "8b6b5f30-0bda-4851-9288-3c3a856e8a6c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1142.053103] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1142.053376] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1142.054939] env[63531]: INFO nova.compute.claims [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] [instance: 4de350e2-55cc-4795-85e6-e58b5b362459] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1142.126811] env[63531]: DEBUG oslo_vmware.api [None req-0043fe24-1de4-469e-b8fa-24bb04ed3c98 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118724, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.298960] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: eb6c84d7-e0e6-4999-9018-889b99dc67ad] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1142.533878] env[63531]: DEBUG oslo_concurrency.lockutils [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Releasing lock "refresh_cache-225e11e5-0558-4ad5-acd4-2a69888fe2d1" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1142.598615] env[63531]: DEBUG nova.network.neutron [req-95199f1c-5707-4708-99ec-3b3df3f1afe6 req-8e6245b0-6fdc-46c3-a17c-d954f4bb994c service nova] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Updated VIF entry in instance network info cache for port baf61e1c-97ff-411c-a0f7-415896653c60. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1142.599075] env[63531]: DEBUG nova.network.neutron [req-95199f1c-5707-4708-99ec-3b3df3f1afe6 req-8e6245b0-6fdc-46c3-a17c-d954f4bb994c service nova] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Updating instance_info_cache with network_info: [{"id": "e9692fb2-f3a9-455c-8317-b533a6af48f7", "address": "fa:16:3e:79:4a:3b", "network": {"id": "caf86428-59a5-4e50-9135-2ab8473af64d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1673994696-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ba8b284854242c392aec5326e996239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape9692fb2-f3", "ovs_interfaceid": "e9692fb2-f3a9-455c-8317-b533a6af48f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "baf61e1c-97ff-411c-a0f7-415896653c60", "address": "fa:16:3e:88:db:88", "network": {"id": "caf86428-59a5-4e50-9135-2ab8473af64d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1673994696-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ba8b284854242c392aec5326e996239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbaf61e1c-97", "ovs_interfaceid": "baf61e1c-97ff-411c-a0f7-415896653c60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1142.625498] env[63531]: DEBUG oslo_vmware.api [None req-0043fe24-1de4-469e-b8fa-24bb04ed3c98 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118724, 'name': ReconfigVM_Task, 'duration_secs': 0.801514} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1142.626008] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0043fe24-1de4-469e-b8fa-24bb04ed3c98 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Releasing lock "ccc8f366-5849-45a6-9745-ff9b2be9c3bb" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1142.626254] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-0043fe24-1de4-469e-b8fa-24bb04ed3c98 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Reconfigured VM to attach interface {{(pid=63531) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 1142.805301] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 531e1852-0c67-4b4d-a0ca-749ac438e79a] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1143.058247] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd1a9f4f-1e12-40d3-b863-93bb3be5df63 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.079833] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8909f440-3013-4a4b-9061-aa4127cd8506 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.086830] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Updating instance '225e11e5-0558-4ad5-acd4-2a69888fe2d1' progress to 83 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1143.101333] env[63531]: DEBUG oslo_concurrency.lockutils [req-95199f1c-5707-4708-99ec-3b3df3f1afe6 req-8e6245b0-6fdc-46c3-a17c-d954f4bb994c service nova] Releasing lock "refresh_cache-ccc8f366-5849-45a6-9745-ff9b2be9c3bb" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1143.130612] env[63531]: DEBUG oslo_concurrency.lockutils [None req-0043fe24-1de4-469e-b8fa-24bb04ed3c98 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "interface-ccc8f366-5849-45a6-9745-ff9b2be9c3bb-baf61e1c-97ff-411c-a0f7-415896653c60" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.141s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1143.171323] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0324d1c-ae26-42f6-8174-49a38f8f9927 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.178939] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ce5cd65-3a22-4687-bc7b-4ac83609323a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.211049] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f79a9f8e-c427-43b9-a2ac-52cf5bd9343b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.218801] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cefcfcd6-14ec-497e-8f83-338b7eb6699a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.232448] env[63531]: DEBUG nova.compute.provider_tree [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1143.308536] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 2dbc4a16-c5a1-47ac-b78e-76ec7d90ca0f] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1143.597028] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1143.597028] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-80206acb-c9e4-4eab-a562-42700c61be62 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.604534] env[63531]: DEBUG oslo_vmware.api [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 1143.604534] env[63531]: value = "task-1118725" [ 1143.604534] env[63531]: _type = "Task" [ 1143.604534] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.612467] env[63531]: DEBUG oslo_vmware.api [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118725, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.735521] env[63531]: DEBUG nova.scheduler.client.report [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1143.811899] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: caf0dee8-cb5a-402b-b533-22a590433f54] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1144.114608] env[63531]: DEBUG oslo_vmware.api [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118725, 'name': PowerOnVM_Task, 'duration_secs': 0.505855} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1144.114959] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1144.115196] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-05e19776-98c8-4414-a038-eca3a5b4823c tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Updating instance '225e11e5-0558-4ad5-acd4-2a69888fe2d1' progress to 100 {{(pid=63531) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1144.240692] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.187s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1144.317115] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 48043b5d-c0d3-4e8a-8d28-5a408d800e11] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1144.701296] env[63531]: DEBUG oslo_concurrency.lockutils [None req-69f0e8a8-b133-497f-923f-76d9773d6e1c tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "interface-ccc8f366-5849-45a6-9745-ff9b2be9c3bb-baf61e1c-97ff-411c-a0f7-415896653c60" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1144.701583] env[63531]: DEBUG oslo_concurrency.lockutils [None req-69f0e8a8-b133-497f-923f-76d9773d6e1c tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "interface-ccc8f366-5849-45a6-9745-ff9b2be9c3bb-baf61e1c-97ff-411c-a0f7-415896653c60" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1144.743527] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Acquiring lock "22989b53-dc06-49ea-9730-886a48ea1e35" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1144.743752] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Lock "22989b53-dc06-49ea-9730-886a48ea1e35" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1144.820298] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 9a3fad46-a15a-451c-bdab-a3c8cc8add07] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1145.204239] env[63531]: DEBUG oslo_concurrency.lockutils [None req-69f0e8a8-b133-497f-923f-76d9773d6e1c tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "ccc8f366-5849-45a6-9745-ff9b2be9c3bb" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1145.204400] env[63531]: DEBUG oslo_concurrency.lockutils [None req-69f0e8a8-b133-497f-923f-76d9773d6e1c tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquired lock "ccc8f366-5849-45a6-9745-ff9b2be9c3bb" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1145.205305] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72955089-8337-4630-a475-528d8df546e8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.223196] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b01bc916-c416-442e-a4ed-a5b4107a5f65 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.248588] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-69f0e8a8-b133-497f-923f-76d9773d6e1c tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Reconfiguring VM to detach interface {{(pid=63531) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 1145.249010] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Lock "22989b53-dc06-49ea-9730-886a48ea1e35" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.505s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1145.249510] env[63531]: DEBUG nova.compute.manager [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] [instance: 4de350e2-55cc-4795-85e6-e58b5b362459] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1145.251924] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-64b15ebc-692d-434e-8e61-a68b3582a844 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.272121] env[63531]: DEBUG oslo_vmware.api [None req-69f0e8a8-b133-497f-923f-76d9773d6e1c tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for the task: (returnval){ [ 1145.272121] env[63531]: value = "task-1118727" [ 1145.272121] env[63531]: _type = "Task" [ 1145.272121] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.279821] env[63531]: DEBUG oslo_vmware.api [None req-69f0e8a8-b133-497f-923f-76d9773d6e1c tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118727, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.323066] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 1e5e28ef-38ce-47be-ac36-8bdf17ab0a47] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1145.766927] env[63531]: DEBUG nova.compute.utils [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1145.768323] env[63531]: DEBUG nova.compute.manager [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] [instance: 4de350e2-55cc-4795-85e6-e58b5b362459] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1145.768637] env[63531]: DEBUG nova.network.neutron [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] [instance: 4de350e2-55cc-4795-85e6-e58b5b362459] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1145.784433] env[63531]: DEBUG oslo_vmware.api [None req-69f0e8a8-b133-497f-923f-76d9773d6e1c tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118727, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.826433] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 0197e03d-6c36-4e73-9472-8e3319eb89ce] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1145.841698] env[63531]: DEBUG nova.policy [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '667e6cf88acd4a6c86ccc99aa7312c6b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b7ddd926b3f64e3e816b11a4e1a02e28', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 1146.112570] env[63531]: DEBUG nova.network.neutron [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] [instance: 4de350e2-55cc-4795-85e6-e58b5b362459] Successfully created port: 85e94a7c-07ba-4f11-b4f0-ac8f6ef24c8e {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1146.275021] env[63531]: DEBUG nova.compute.manager [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] [instance: 4de350e2-55cc-4795-85e6-e58b5b362459] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1146.286200] env[63531]: DEBUG oslo_vmware.api [None req-69f0e8a8-b133-497f-923f-76d9773d6e1c tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118727, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.331237] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 64fb5831-0789-482b-bb98-67b29868c4c9] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1146.785912] env[63531]: DEBUG oslo_vmware.api [None req-69f0e8a8-b133-497f-923f-76d9773d6e1c tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118727, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.833294] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 2998fb3d-6677-4fb3-952a-c74d8c6b7b9f] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1147.000929] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3447c1a3-dd26-481f-97eb-b24f05a0cc18 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "225e11e5-0558-4ad5-acd4-2a69888fe2d1" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1147.001223] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3447c1a3-dd26-481f-97eb-b24f05a0cc18 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "225e11e5-0558-4ad5-acd4-2a69888fe2d1" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1147.001413] env[63531]: DEBUG nova.compute.manager [None req-3447c1a3-dd26-481f-97eb-b24f05a0cc18 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Going to confirm migration 9 {{(pid=63531) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1147.284375] env[63531]: DEBUG oslo_vmware.api [None req-69f0e8a8-b133-497f-923f-76d9773d6e1c tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118727, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.287627] env[63531]: DEBUG nova.compute.manager [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] [instance: 4de350e2-55cc-4795-85e6-e58b5b362459] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1147.312848] env[63531]: DEBUG nova.virt.hardware [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1147.313110] env[63531]: DEBUG nova.virt.hardware [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1147.313279] env[63531]: DEBUG nova.virt.hardware [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1147.313470] env[63531]: DEBUG nova.virt.hardware [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1147.313624] env[63531]: DEBUG nova.virt.hardware [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1147.313778] env[63531]: DEBUG nova.virt.hardware [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1147.313988] env[63531]: DEBUG nova.virt.hardware [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1147.314166] env[63531]: DEBUG nova.virt.hardware [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1147.314392] env[63531]: DEBUG nova.virt.hardware [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1147.314568] env[63531]: DEBUG nova.virt.hardware [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1147.314803] env[63531]: DEBUG nova.virt.hardware [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1147.316073] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e72c2d5b-a449-40a9-b37b-e16ed5a8561d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.323401] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b840bc7f-8875-4e07-9052-fb2e0611ba80 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.336250] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 29dd6de3-2765-408c-acc0-da47e5e0a977] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1147.566061] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3447c1a3-dd26-481f-97eb-b24f05a0cc18 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "refresh_cache-225e11e5-0558-4ad5-acd4-2a69888fe2d1" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1147.566276] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3447c1a3-dd26-481f-97eb-b24f05a0cc18 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquired lock "refresh_cache-225e11e5-0558-4ad5-acd4-2a69888fe2d1" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1147.566459] env[63531]: DEBUG nova.network.neutron [None req-3447c1a3-dd26-481f-97eb-b24f05a0cc18 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1147.566820] env[63531]: DEBUG nova.objects.instance [None req-3447c1a3-dd26-481f-97eb-b24f05a0cc18 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lazy-loading 'info_cache' on Instance uuid 225e11e5-0558-4ad5-acd4-2a69888fe2d1 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1147.654717] env[63531]: DEBUG nova.network.neutron [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] [instance: 4de350e2-55cc-4795-85e6-e58b5b362459] Successfully updated port: 85e94a7c-07ba-4f11-b4f0-ac8f6ef24c8e {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1147.662861] env[63531]: DEBUG nova.compute.manager [req-ebea43c0-5f94-4e30-bdd1-f9031b1f251a req-c69ab230-034d-4806-a376-27a322c4a2e5 service nova] [instance: 4de350e2-55cc-4795-85e6-e58b5b362459] Received event network-vif-plugged-85e94a7c-07ba-4f11-b4f0-ac8f6ef24c8e {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1147.662861] env[63531]: DEBUG oslo_concurrency.lockutils [req-ebea43c0-5f94-4e30-bdd1-f9031b1f251a req-c69ab230-034d-4806-a376-27a322c4a2e5 service nova] Acquiring lock "4de350e2-55cc-4795-85e6-e58b5b362459-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1147.663058] env[63531]: DEBUG oslo_concurrency.lockutils [req-ebea43c0-5f94-4e30-bdd1-f9031b1f251a req-c69ab230-034d-4806-a376-27a322c4a2e5 service nova] Lock "4de350e2-55cc-4795-85e6-e58b5b362459-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1147.663201] env[63531]: DEBUG oslo_concurrency.lockutils [req-ebea43c0-5f94-4e30-bdd1-f9031b1f251a req-c69ab230-034d-4806-a376-27a322c4a2e5 service nova] Lock "4de350e2-55cc-4795-85e6-e58b5b362459-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1147.663337] env[63531]: DEBUG nova.compute.manager [req-ebea43c0-5f94-4e30-bdd1-f9031b1f251a req-c69ab230-034d-4806-a376-27a322c4a2e5 service nova] [instance: 4de350e2-55cc-4795-85e6-e58b5b362459] No waiting events found dispatching network-vif-plugged-85e94a7c-07ba-4f11-b4f0-ac8f6ef24c8e {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1147.663557] env[63531]: WARNING nova.compute.manager [req-ebea43c0-5f94-4e30-bdd1-f9031b1f251a req-c69ab230-034d-4806-a376-27a322c4a2e5 service nova] [instance: 4de350e2-55cc-4795-85e6-e58b5b362459] Received unexpected event network-vif-plugged-85e94a7c-07ba-4f11-b4f0-ac8f6ef24c8e for instance with vm_state building and task_state spawning. [ 1147.788017] env[63531]: DEBUG oslo_vmware.api [None req-69f0e8a8-b133-497f-923f-76d9773d6e1c tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118727, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.841596] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: e7fb00a3-0d2e-4f54-950d-337307112d7a] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1148.157437] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Acquiring lock "refresh_cache-4de350e2-55cc-4795-85e6-e58b5b362459" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1148.157697] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Acquired lock "refresh_cache-4de350e2-55cc-4795-85e6-e58b5b362459" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1148.157916] env[63531]: DEBUG nova.network.neutron [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] [instance: 4de350e2-55cc-4795-85e6-e58b5b362459] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1148.286536] env[63531]: DEBUG oslo_vmware.api [None req-69f0e8a8-b133-497f-923f-76d9773d6e1c tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118727, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.344814] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 1ca7b851-2a06-4181-8271-58aafcd322d6] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1148.723340] env[63531]: DEBUG nova.network.neutron [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] [instance: 4de350e2-55cc-4795-85e6-e58b5b362459] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1148.794373] env[63531]: DEBUG oslo_vmware.api [None req-69f0e8a8-b133-497f-923f-76d9773d6e1c tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118727, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.849821] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 5e9042b3-4dc2-4fa3-a664-c4b49a22e400] Instance has had 0 of 5 cleanup attempts {{(pid=63531) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1148.891824] env[63531]: DEBUG nova.network.neutron [None req-3447c1a3-dd26-481f-97eb-b24f05a0cc18 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Updating instance_info_cache with network_info: [{"id": "8b6b5f30-0bda-4851-9288-3c3a856e8a6c", "address": "fa:16:3e:8d:e8:d4", "network": {"id": "ea2d47d3-a502-4afd-9bc5-ee4a37017901", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1361282648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94a95667c00947ea9183a6307c569c90", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6c45fd-e930-495a-9cb7-df84eda443b1", "external-id": "nsx-vlan-transportzone-407", "segmentation_id": 407, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8b6b5f30-0b", "ovs_interfaceid": "8b6b5f30-0bda-4851-9288-3c3a856e8a6c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1148.898083] env[63531]: DEBUG nova.network.neutron [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] [instance: 4de350e2-55cc-4795-85e6-e58b5b362459] Updating instance_info_cache with network_info: [{"id": "85e94a7c-07ba-4f11-b4f0-ac8f6ef24c8e", "address": "fa:16:3e:6e:ff:85", "network": {"id": "d489dac0-e15f-488a-85e3-eeb613ec3f4e", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1224268557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b7ddd926b3f64e3e816b11a4e1a02e28", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "09bf081b-cdf0-4977-abe2-2339a87409ab", "external-id": "nsx-vlan-transportzone-378", "segmentation_id": 378, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap85e94a7c-07", "ovs_interfaceid": "85e94a7c-07ba-4f11-b4f0-ac8f6ef24c8e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1149.288977] env[63531]: DEBUG oslo_vmware.api [None req-69f0e8a8-b133-497f-923f-76d9773d6e1c tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118727, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.353655] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1149.353918] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Cleaning up deleted instances with incomplete migration {{(pid=63531) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 1149.394982] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3447c1a3-dd26-481f-97eb-b24f05a0cc18 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Releasing lock "refresh_cache-225e11e5-0558-4ad5-acd4-2a69888fe2d1" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1149.395307] env[63531]: DEBUG nova.objects.instance [None req-3447c1a3-dd26-481f-97eb-b24f05a0cc18 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lazy-loading 'migration_context' on Instance uuid 225e11e5-0558-4ad5-acd4-2a69888fe2d1 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1149.400317] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Releasing lock "refresh_cache-4de350e2-55cc-4795-85e6-e58b5b362459" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1149.400664] env[63531]: DEBUG nova.compute.manager [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] [instance: 4de350e2-55cc-4795-85e6-e58b5b362459] Instance network_info: |[{"id": "85e94a7c-07ba-4f11-b4f0-ac8f6ef24c8e", "address": "fa:16:3e:6e:ff:85", "network": {"id": "d489dac0-e15f-488a-85e3-eeb613ec3f4e", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1224268557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b7ddd926b3f64e3e816b11a4e1a02e28", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "09bf081b-cdf0-4977-abe2-2339a87409ab", "external-id": "nsx-vlan-transportzone-378", "segmentation_id": 378, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap85e94a7c-07", "ovs_interfaceid": "85e94a7c-07ba-4f11-b4f0-ac8f6ef24c8e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1149.401113] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] [instance: 4de350e2-55cc-4795-85e6-e58b5b362459] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6e:ff:85', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '09bf081b-cdf0-4977-abe2-2339a87409ab', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '85e94a7c-07ba-4f11-b4f0-ac8f6ef24c8e', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1149.408385] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Creating folder: Project (b7ddd926b3f64e3e816b11a4e1a02e28). Parent ref: group-v244585. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1149.409174] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bcab1533-1ce4-4f58-8e53-5923314a5dfd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.420240] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Created folder: Project (b7ddd926b3f64e3e816b11a4e1a02e28) in parent group-v244585. [ 1149.420427] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Creating folder: Instances. Parent ref: group-v244816. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1149.420680] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-40b6173e-4709-4b06-ae49-7ea0cd1f9573 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.429562] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Created folder: Instances in parent group-v244816. [ 1149.429766] env[63531]: DEBUG oslo.service.loopingcall [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1149.429955] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4de350e2-55cc-4795-85e6-e58b5b362459] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1149.430404] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-493ca54f-a1bc-42ec-8876-1939211c39b7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.449152] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1149.449152] env[63531]: value = "task-1118730" [ 1149.449152] env[63531]: _type = "Task" [ 1149.449152] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.462258] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118730, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.695630] env[63531]: DEBUG nova.compute.manager [req-cd60c78a-3ba3-4fa0-8317-22b13f65a5c9 req-8ddf0ecf-dbdc-45f2-9108-2f4cc0fe3f0b service nova] [instance: 4de350e2-55cc-4795-85e6-e58b5b362459] Received event network-changed-85e94a7c-07ba-4f11-b4f0-ac8f6ef24c8e {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1149.695862] env[63531]: DEBUG nova.compute.manager [req-cd60c78a-3ba3-4fa0-8317-22b13f65a5c9 req-8ddf0ecf-dbdc-45f2-9108-2f4cc0fe3f0b service nova] [instance: 4de350e2-55cc-4795-85e6-e58b5b362459] Refreshing instance network info cache due to event network-changed-85e94a7c-07ba-4f11-b4f0-ac8f6ef24c8e. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1149.696074] env[63531]: DEBUG oslo_concurrency.lockutils [req-cd60c78a-3ba3-4fa0-8317-22b13f65a5c9 req-8ddf0ecf-dbdc-45f2-9108-2f4cc0fe3f0b service nova] Acquiring lock "refresh_cache-4de350e2-55cc-4795-85e6-e58b5b362459" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1149.696230] env[63531]: DEBUG oslo_concurrency.lockutils [req-cd60c78a-3ba3-4fa0-8317-22b13f65a5c9 req-8ddf0ecf-dbdc-45f2-9108-2f4cc0fe3f0b service nova] Acquired lock "refresh_cache-4de350e2-55cc-4795-85e6-e58b5b362459" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1149.696438] env[63531]: DEBUG nova.network.neutron [req-cd60c78a-3ba3-4fa0-8317-22b13f65a5c9 req-8ddf0ecf-dbdc-45f2-9108-2f4cc0fe3f0b service nova] [instance: 4de350e2-55cc-4795-85e6-e58b5b362459] Refreshing network info cache for port 85e94a7c-07ba-4f11-b4f0-ac8f6ef24c8e {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1149.789578] env[63531]: DEBUG oslo_vmware.api [None req-69f0e8a8-b133-497f-923f-76d9773d6e1c tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118727, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.856575] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1149.898727] env[63531]: DEBUG nova.objects.base [None req-3447c1a3-dd26-481f-97eb-b24f05a0cc18 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Object Instance<225e11e5-0558-4ad5-acd4-2a69888fe2d1> lazy-loaded attributes: info_cache,migration_context {{(pid=63531) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1149.900418] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6310fef0-3fc3-4118-8a80-1bcadc745b69 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.919175] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4accb7a3-0be6-4892-95f8-95849813a8e9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.931187] env[63531]: DEBUG oslo_vmware.api [None req-3447c1a3-dd26-481f-97eb-b24f05a0cc18 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 1149.931187] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]529bc0b6-c61b-abcd-ca01-90415e3f7336" [ 1149.931187] env[63531]: _type = "Task" [ 1149.931187] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.939966] env[63531]: DEBUG oslo_vmware.api [None req-3447c1a3-dd26-481f-97eb-b24f05a0cc18 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]529bc0b6-c61b-abcd-ca01-90415e3f7336, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.958593] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118730, 'name': CreateVM_Task, 'duration_secs': 0.454481} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.959724] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4de350e2-55cc-4795-85e6-e58b5b362459] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1149.959724] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1149.959900] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1149.960193] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1149.960439] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4ca084da-1cf2-4388-b91e-870434f4225c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.964931] env[63531]: DEBUG oslo_vmware.api [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Waiting for the task: (returnval){ [ 1149.964931] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52b3404b-8107-38f2-611c-9f9c8f78cd3b" [ 1149.964931] env[63531]: _type = "Task" [ 1149.964931] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.972246] env[63531]: DEBUG oslo_vmware.api [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52b3404b-8107-38f2-611c-9f9c8f78cd3b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.292118] env[63531]: DEBUG oslo_vmware.api [None req-69f0e8a8-b133-497f-923f-76d9773d6e1c tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118727, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.399806] env[63531]: DEBUG nova.network.neutron [req-cd60c78a-3ba3-4fa0-8317-22b13f65a5c9 req-8ddf0ecf-dbdc-45f2-9108-2f4cc0fe3f0b service nova] [instance: 4de350e2-55cc-4795-85e6-e58b5b362459] Updated VIF entry in instance network info cache for port 85e94a7c-07ba-4f11-b4f0-ac8f6ef24c8e. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1150.400382] env[63531]: DEBUG nova.network.neutron [req-cd60c78a-3ba3-4fa0-8317-22b13f65a5c9 req-8ddf0ecf-dbdc-45f2-9108-2f4cc0fe3f0b service nova] [instance: 4de350e2-55cc-4795-85e6-e58b5b362459] Updating instance_info_cache with network_info: [{"id": "85e94a7c-07ba-4f11-b4f0-ac8f6ef24c8e", "address": "fa:16:3e:6e:ff:85", "network": {"id": "d489dac0-e15f-488a-85e3-eeb613ec3f4e", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1224268557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b7ddd926b3f64e3e816b11a4e1a02e28", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "09bf081b-cdf0-4977-abe2-2339a87409ab", "external-id": "nsx-vlan-transportzone-378", "segmentation_id": 378, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap85e94a7c-07", "ovs_interfaceid": "85e94a7c-07ba-4f11-b4f0-ac8f6ef24c8e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1150.441493] env[63531]: DEBUG oslo_vmware.api [None req-3447c1a3-dd26-481f-97eb-b24f05a0cc18 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]529bc0b6-c61b-abcd-ca01-90415e3f7336, 'name': SearchDatastore_Task, 'duration_secs': 0.025231} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.441829] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3447c1a3-dd26-481f-97eb-b24f05a0cc18 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1150.442089] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3447c1a3-dd26-481f-97eb-b24f05a0cc18 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1150.475339] env[63531]: DEBUG oslo_vmware.api [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52b3404b-8107-38f2-611c-9f9c8f78cd3b, 'name': SearchDatastore_Task, 'duration_secs': 0.029696} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.475631] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1150.475873] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] [instance: 4de350e2-55cc-4795-85e6-e58b5b362459] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1150.476130] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1150.476282] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1150.476463] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1150.476725] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2715a902-ac42-46a5-b156-0fa17c082761 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.485273] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1150.485462] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1150.486215] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1b3de38f-af8e-4e52-ba4a-cc2093402c33 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.491784] env[63531]: DEBUG oslo_vmware.api [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Waiting for the task: (returnval){ [ 1150.491784] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]521f0783-01dd-1c04-8bf3-c00f2f0f5936" [ 1150.491784] env[63531]: _type = "Task" [ 1150.491784] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.500998] env[63531]: DEBUG oslo_vmware.api [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]521f0783-01dd-1c04-8bf3-c00f2f0f5936, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.794324] env[63531]: DEBUG oslo_vmware.api [None req-69f0e8a8-b133-497f-923f-76d9773d6e1c tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118727, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.903619] env[63531]: DEBUG oslo_concurrency.lockutils [req-cd60c78a-3ba3-4fa0-8317-22b13f65a5c9 req-8ddf0ecf-dbdc-45f2-9108-2f4cc0fe3f0b service nova] Releasing lock "refresh_cache-4de350e2-55cc-4795-85e6-e58b5b362459" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1151.006045] env[63531]: DEBUG oslo_vmware.api [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]521f0783-01dd-1c04-8bf3-c00f2f0f5936, 'name': SearchDatastore_Task, 'duration_secs': 0.014055} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.009877] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f4076c7d-a857-48fb-a2b9-89a5b508f2cf {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.016750] env[63531]: DEBUG oslo_vmware.api [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Waiting for the task: (returnval){ [ 1151.016750] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52394f88-33e3-61e2-43d6-1b384b464613" [ 1151.016750] env[63531]: _type = "Task" [ 1151.016750] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.029194] env[63531]: DEBUG oslo_vmware.api [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52394f88-33e3-61e2-43d6-1b384b464613, 'name': SearchDatastore_Task} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.030288] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1151.030288] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 4de350e2-55cc-4795-85e6-e58b5b362459/4de350e2-55cc-4795-85e6-e58b5b362459.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1151.030288] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c7bc56e7-382d-487d-9f73-9c1f4b797279 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.036727] env[63531]: DEBUG oslo_vmware.api [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Waiting for the task: (returnval){ [ 1151.036727] env[63531]: value = "task-1118731" [ 1151.036727] env[63531]: _type = "Task" [ 1151.036727] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.040783] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1151.049279] env[63531]: DEBUG oslo_vmware.api [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Task: {'id': task-1118731, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.071372] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c4fb046-f419-4eaf-acef-7289ce6a09d9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.079774] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6d6fdb3-52fc-4771-855c-4affc894d8a7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.113573] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a350a98-37a3-4fee-9bd9-4cd2a66b897f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.121274] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c473a65f-46c7-4298-9ba6-30002767e09f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.135220] env[63531]: DEBUG nova.compute.provider_tree [None req-3447c1a3-dd26-481f-97eb-b24f05a0cc18 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1151.296457] env[63531]: DEBUG oslo_vmware.api [None req-69f0e8a8-b133-497f-923f-76d9773d6e1c tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118727, 'name': ReconfigVM_Task, 'duration_secs': 5.805895} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.296735] env[63531]: DEBUG oslo_concurrency.lockutils [None req-69f0e8a8-b133-497f-923f-76d9773d6e1c tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Releasing lock "ccc8f366-5849-45a6-9745-ff9b2be9c3bb" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1151.296978] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-69f0e8a8-b133-497f-923f-76d9773d6e1c tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Reconfigured VM to detach interface {{(pid=63531) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 1151.548854] env[63531]: DEBUG oslo_vmware.api [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Task: {'id': task-1118731, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.487589} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.549373] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 4de350e2-55cc-4795-85e6-e58b5b362459/4de350e2-55cc-4795-85e6-e58b5b362459.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1151.549517] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] [instance: 4de350e2-55cc-4795-85e6-e58b5b362459] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1151.549689] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1151.551615] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dc2230ab-9397-4758-b961-60794ee7322c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.552258] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._sync_power_states {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1151.557912] env[63531]: DEBUG oslo_vmware.api [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Waiting for the task: (returnval){ [ 1151.557912] env[63531]: value = "task-1118732" [ 1151.557912] env[63531]: _type = "Task" [ 1151.557912] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.566507] env[63531]: DEBUG oslo_vmware.api [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Task: {'id': task-1118732, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.638033] env[63531]: DEBUG nova.scheduler.client.report [None req-3447c1a3-dd26-481f-97eb-b24f05a0cc18 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1152.055654] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Getting list of instances from cluster (obj){ [ 1152.055654] env[63531]: value = "domain-c8" [ 1152.055654] env[63531]: _type = "ClusterComputeResource" [ 1152.055654] env[63531]: } {{(pid=63531) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 1152.056741] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d46ceff7-1025-4d81-873b-4439a7cce66d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.069183] env[63531]: DEBUG oslo_vmware.api [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Task: {'id': task-1118732, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.261891} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.076018] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] [instance: 4de350e2-55cc-4795-85e6-e58b5b362459] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1152.076313] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Got total of 6 instances {{(pid=63531) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 1152.076466] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Triggering sync for uuid ccc8f366-5849-45a6-9745-ff9b2be9c3bb {{(pid=63531) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1152.076651] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Triggering sync for uuid 98853767-791e-4568-8bb8-30dfcc473ee3 {{(pid=63531) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1152.076805] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Triggering sync for uuid 225e11e5-0558-4ad5-acd4-2a69888fe2d1 {{(pid=63531) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1152.076953] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Triggering sync for uuid 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97 {{(pid=63531) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1152.077113] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Triggering sync for uuid 2d51997d-f88c-468c-9a56-0331a06ca56c {{(pid=63531) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1152.077567] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Triggering sync for uuid 4de350e2-55cc-4795-85e6-e58b5b362459 {{(pid=63531) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1152.077988] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcb74f63-f0ae-42e5-8395-1136e2226434 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.080322] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Acquiring lock "ccc8f366-5849-45a6-9745-ff9b2be9c3bb" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1152.080543] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Lock "ccc8f366-5849-45a6-9745-ff9b2be9c3bb" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1152.080822] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Acquiring lock "98853767-791e-4568-8bb8-30dfcc473ee3" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1152.081020] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Lock "98853767-791e-4568-8bb8-30dfcc473ee3" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1152.081266] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Acquiring lock "225e11e5-0558-4ad5-acd4-2a69888fe2d1" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1152.081474] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Acquiring lock "3794f44c-8f6d-4bab-b9e0-9f6c819e1a97" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1152.081664] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Lock "3794f44c-8f6d-4bab-b9e0-9f6c819e1a97" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1152.081902] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Acquiring lock "2d51997d-f88c-468c-9a56-0331a06ca56c" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1152.082093] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Lock "2d51997d-f88c-468c-9a56-0331a06ca56c" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1152.082328] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Acquiring lock "4de350e2-55cc-4795-85e6-e58b5b362459" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1152.083248] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76d81b07-1e23-454e-9298-cd2f270180b8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.085813] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00f16550-d18f-4946-9ae3-89b19bf00136 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.088472] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0f60948-6082-42de-8c10-471e3ec739db {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.091014] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3907af93-0451-4e5c-bc38-ff239ca1952b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.124998] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] [instance: 4de350e2-55cc-4795-85e6-e58b5b362459] Reconfiguring VM instance instance-00000074 to attach disk [datastore1] 4de350e2-55cc-4795-85e6-e58b5b362459/4de350e2-55cc-4795-85e6-e58b5b362459.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1152.126475] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-46d38c3e-4669-405b-b2f5-c59f443af9aa {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.153267] env[63531]: DEBUG oslo_vmware.api [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Waiting for the task: (returnval){ [ 1152.153267] env[63531]: value = "task-1118733" [ 1152.153267] env[63531]: _type = "Task" [ 1152.153267] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.170203] env[63531]: DEBUG oslo_vmware.api [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Task: {'id': task-1118733, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.626541] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Lock "ccc8f366-5849-45a6-9745-ff9b2be9c3bb" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.546s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1152.627397] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Lock "3794f44c-8f6d-4bab-b9e0-9f6c819e1a97" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.546s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1152.627833] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Lock "98853767-791e-4568-8bb8-30dfcc473ee3" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.547s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1152.648474] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Lock "2d51997d-f88c-468c-9a56-0331a06ca56c" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.566s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1152.651986] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3447c1a3-dd26-481f-97eb-b24f05a0cc18 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.210s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1152.654850] env[63531]: DEBUG oslo_concurrency.lockutils [None req-69f0e8a8-b133-497f-923f-76d9773d6e1c tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "refresh_cache-ccc8f366-5849-45a6-9745-ff9b2be9c3bb" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1152.655029] env[63531]: DEBUG oslo_concurrency.lockutils [None req-69f0e8a8-b133-497f-923f-76d9773d6e1c tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquired lock "refresh_cache-ccc8f366-5849-45a6-9745-ff9b2be9c3bb" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1152.655210] env[63531]: DEBUG nova.network.neutron [None req-69f0e8a8-b133-497f-923f-76d9773d6e1c tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1152.666615] env[63531]: DEBUG oslo_vmware.api [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Task: {'id': task-1118733, 'name': ReconfigVM_Task, 'duration_secs': 0.321468} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.667356] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] [instance: 4de350e2-55cc-4795-85e6-e58b5b362459] Reconfigured VM instance instance-00000074 to attach disk [datastore1] 4de350e2-55cc-4795-85e6-e58b5b362459/4de350e2-55cc-4795-85e6-e58b5b362459.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1152.667545] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ada670d3-d2c9-4b60-98a5-bf8ad1b8b95d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.673961] env[63531]: DEBUG oslo_vmware.api [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Waiting for the task: (returnval){ [ 1152.673961] env[63531]: value = "task-1118734" [ 1152.673961] env[63531]: _type = "Task" [ 1152.673961] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.682836] env[63531]: DEBUG oslo_vmware.api [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Task: {'id': task-1118734, 'name': Rename_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.191020] env[63531]: DEBUG oslo_vmware.api [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Task: {'id': task-1118734, 'name': Rename_Task, 'duration_secs': 0.145426} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.191020] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] [instance: 4de350e2-55cc-4795-85e6-e58b5b362459] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1153.191020] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-52c8a2d9-6807-4c50-a7ff-21745d3df5a4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.196168] env[63531]: DEBUG oslo_vmware.api [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Waiting for the task: (returnval){ [ 1153.196168] env[63531]: value = "task-1118735" [ 1153.196168] env[63531]: _type = "Task" [ 1153.196168] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.205210] env[63531]: DEBUG oslo_vmware.api [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Task: {'id': task-1118735, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.220142] env[63531]: INFO nova.scheduler.client.report [None req-3447c1a3-dd26-481f-97eb-b24f05a0cc18 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Deleted allocation for migration b80d81d1-28a5-4648-9e17-d1ccbade2da2 [ 1153.411552] env[63531]: DEBUG nova.compute.manager [req-592095c0-5b8c-4f80-b84e-48950d64766b req-7e968be8-ad3f-415d-a460-da00d0878b2b service nova] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Received event network-changed-e9692fb2-f3a9-455c-8317-b533a6af48f7 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1153.411552] env[63531]: DEBUG nova.compute.manager [req-592095c0-5b8c-4f80-b84e-48950d64766b req-7e968be8-ad3f-415d-a460-da00d0878b2b service nova] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Refreshing instance network info cache due to event network-changed-e9692fb2-f3a9-455c-8317-b533a6af48f7. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1153.411648] env[63531]: DEBUG oslo_concurrency.lockutils [req-592095c0-5b8c-4f80-b84e-48950d64766b req-7e968be8-ad3f-415d-a460-da00d0878b2b service nova] Acquiring lock "refresh_cache-ccc8f366-5849-45a6-9745-ff9b2be9c3bb" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1153.426181] env[63531]: INFO nova.network.neutron [None req-69f0e8a8-b133-497f-923f-76d9773d6e1c tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Port baf61e1c-97ff-411c-a0f7-415896653c60 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1153.426578] env[63531]: DEBUG nova.network.neutron [None req-69f0e8a8-b133-497f-923f-76d9773d6e1c tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Updating instance_info_cache with network_info: [{"id": "e9692fb2-f3a9-455c-8317-b533a6af48f7", "address": "fa:16:3e:79:4a:3b", "network": {"id": "caf86428-59a5-4e50-9135-2ab8473af64d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1673994696-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ba8b284854242c392aec5326e996239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape9692fb2-f3", "ovs_interfaceid": "e9692fb2-f3a9-455c-8317-b533a6af48f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1153.706084] env[63531]: DEBUG oslo_vmware.api [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Task: {'id': task-1118735, 'name': PowerOnVM_Task, 'duration_secs': 0.4299} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.706374] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] [instance: 4de350e2-55cc-4795-85e6-e58b5b362459] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1153.706580] env[63531]: INFO nova.compute.manager [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] [instance: 4de350e2-55cc-4795-85e6-e58b5b362459] Took 6.42 seconds to spawn the instance on the hypervisor. [ 1153.706760] env[63531]: DEBUG nova.compute.manager [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] [instance: 4de350e2-55cc-4795-85e6-e58b5b362459] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1153.707528] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b540dde-eacf-4840-9b19-254f7cb34b2d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.726795] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3447c1a3-dd26-481f-97eb-b24f05a0cc18 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "225e11e5-0558-4ad5-acd4-2a69888fe2d1" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.725s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1153.727748] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Lock "225e11e5-0558-4ad5-acd4-2a69888fe2d1" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 1.646s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1153.727945] env[63531]: INFO nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] During sync_power_state the instance has a pending task (deleting). Skip. [ 1153.728151] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Lock "225e11e5-0558-4ad5-acd4-2a69888fe2d1" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1153.929024] env[63531]: DEBUG oslo_concurrency.lockutils [None req-69f0e8a8-b133-497f-923f-76d9773d6e1c tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Releasing lock "refresh_cache-ccc8f366-5849-45a6-9745-ff9b2be9c3bb" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1153.931523] env[63531]: DEBUG oslo_concurrency.lockutils [req-592095c0-5b8c-4f80-b84e-48950d64766b req-7e968be8-ad3f-415d-a460-da00d0878b2b service nova] Acquired lock "refresh_cache-ccc8f366-5849-45a6-9745-ff9b2be9c3bb" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1153.931768] env[63531]: DEBUG nova.network.neutron [req-592095c0-5b8c-4f80-b84e-48950d64766b req-7e968be8-ad3f-415d-a460-da00d0878b2b service nova] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Refreshing network info cache for port e9692fb2-f3a9-455c-8317-b533a6af48f7 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1153.979941] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5eadccc3-df04-41f8-b336-4b32353995e2 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "interface-98853767-791e-4568-8bb8-30dfcc473ee3-baf61e1c-97ff-411c-a0f7-415896653c60" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1153.980184] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5eadccc3-df04-41f8-b336-4b32353995e2 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "interface-98853767-791e-4568-8bb8-30dfcc473ee3-baf61e1c-97ff-411c-a0f7-415896653c60" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1153.980609] env[63531]: DEBUG nova.objects.instance [None req-5eadccc3-df04-41f8-b336-4b32353995e2 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lazy-loading 'flavor' on Instance uuid 98853767-791e-4568-8bb8-30dfcc473ee3 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1154.224804] env[63531]: INFO nova.compute.manager [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] [instance: 4de350e2-55cc-4795-85e6-e58b5b362459] Took 12.19 seconds to build instance. [ 1154.434951] env[63531]: DEBUG oslo_concurrency.lockutils [None req-69f0e8a8-b133-497f-923f-76d9773d6e1c tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "interface-ccc8f366-5849-45a6-9745-ff9b2be9c3bb-baf61e1c-97ff-411c-a0f7-415896653c60" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.733s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1154.581154] env[63531]: DEBUG nova.objects.instance [None req-5eadccc3-df04-41f8-b336-4b32353995e2 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lazy-loading 'pci_requests' on Instance uuid 98853767-791e-4568-8bb8-30dfcc473ee3 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1154.687016] env[63531]: DEBUG nova.network.neutron [req-592095c0-5b8c-4f80-b84e-48950d64766b req-7e968be8-ad3f-415d-a460-da00d0878b2b service nova] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Updated VIF entry in instance network info cache for port e9692fb2-f3a9-455c-8317-b533a6af48f7. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1154.687496] env[63531]: DEBUG nova.network.neutron [req-592095c0-5b8c-4f80-b84e-48950d64766b req-7e968be8-ad3f-415d-a460-da00d0878b2b service nova] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Updating instance_info_cache with network_info: [{"id": "e9692fb2-f3a9-455c-8317-b533a6af48f7", "address": "fa:16:3e:79:4a:3b", "network": {"id": "caf86428-59a5-4e50-9135-2ab8473af64d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1673994696-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ba8b284854242c392aec5326e996239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape9692fb2-f3", "ovs_interfaceid": "e9692fb2-f3a9-455c-8317-b533a6af48f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1154.727164] env[63531]: DEBUG oslo_concurrency.lockutils [None req-7ea85ff7-4808-44a0-9740-dfaf7cffbb03 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Lock "4de350e2-55cc-4795-85e6-e58b5b362459" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.700s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1154.727507] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Lock "4de350e2-55cc-4795-85e6-e58b5b362459" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 2.645s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1154.727744] env[63531]: INFO nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] [instance: 4de350e2-55cc-4795-85e6-e58b5b362459] During sync_power_state the instance has a pending task (spawning). Skip. [ 1154.727974] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Lock "4de350e2-55cc-4795-85e6-e58b5b362459" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1155.026789] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3447c1a3-dd26-481f-97eb-b24f05a0cc18 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "225e11e5-0558-4ad5-acd4-2a69888fe2d1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1155.027176] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3447c1a3-dd26-481f-97eb-b24f05a0cc18 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "225e11e5-0558-4ad5-acd4-2a69888fe2d1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1155.027343] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3447c1a3-dd26-481f-97eb-b24f05a0cc18 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "225e11e5-0558-4ad5-acd4-2a69888fe2d1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1155.027685] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3447c1a3-dd26-481f-97eb-b24f05a0cc18 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "225e11e5-0558-4ad5-acd4-2a69888fe2d1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1155.028011] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3447c1a3-dd26-481f-97eb-b24f05a0cc18 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "225e11e5-0558-4ad5-acd4-2a69888fe2d1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1155.030909] env[63531]: INFO nova.compute.manager [None req-3447c1a3-dd26-481f-97eb-b24f05a0cc18 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Terminating instance [ 1155.033084] env[63531]: DEBUG nova.compute.manager [None req-3447c1a3-dd26-481f-97eb-b24f05a0cc18 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1155.033240] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-3447c1a3-dd26-481f-97eb-b24f05a0cc18 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1155.034135] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-710644ae-6d85-4722-8acb-0aea019fdd9f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.042896] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-3447c1a3-dd26-481f-97eb-b24f05a0cc18 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1155.043404] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1095df65-218f-4d28-8164-7b77307e1359 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.049587] env[63531]: DEBUG oslo_vmware.api [None req-3447c1a3-dd26-481f-97eb-b24f05a0cc18 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 1155.049587] env[63531]: value = "task-1118736" [ 1155.049587] env[63531]: _type = "Task" [ 1155.049587] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.057216] env[63531]: DEBUG oslo_vmware.api [None req-3447c1a3-dd26-481f-97eb-b24f05a0cc18 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118736, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.084226] env[63531]: DEBUG nova.objects.base [None req-5eadccc3-df04-41f8-b336-4b32353995e2 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Object Instance<98853767-791e-4568-8bb8-30dfcc473ee3> lazy-loaded attributes: flavor,pci_requests {{(pid=63531) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1155.084480] env[63531]: DEBUG nova.network.neutron [None req-5eadccc3-df04-41f8-b336-4b32353995e2 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1155.160549] env[63531]: DEBUG nova.policy [None req-5eadccc3-df04-41f8-b336-4b32353995e2 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '36cb2f0de54f417384fa1fb2210de4e4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8ba8b284854242c392aec5326e996239', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 1155.190364] env[63531]: DEBUG oslo_concurrency.lockutils [req-592095c0-5b8c-4f80-b84e-48950d64766b req-7e968be8-ad3f-415d-a460-da00d0878b2b service nova] Releasing lock "refresh_cache-ccc8f366-5849-45a6-9745-ff9b2be9c3bb" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1155.405523] env[63531]: DEBUG oslo_concurrency.lockutils [None req-379abbe1-61a6-4c63-9167-2135a9f50e91 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Acquiring lock "4de350e2-55cc-4795-85e6-e58b5b362459" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1155.405811] env[63531]: DEBUG oslo_concurrency.lockutils [None req-379abbe1-61a6-4c63-9167-2135a9f50e91 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Lock "4de350e2-55cc-4795-85e6-e58b5b362459" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1155.406035] env[63531]: DEBUG oslo_concurrency.lockutils [None req-379abbe1-61a6-4c63-9167-2135a9f50e91 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Acquiring lock "4de350e2-55cc-4795-85e6-e58b5b362459-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1155.406233] env[63531]: DEBUG oslo_concurrency.lockutils [None req-379abbe1-61a6-4c63-9167-2135a9f50e91 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Lock "4de350e2-55cc-4795-85e6-e58b5b362459-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1155.406407] env[63531]: DEBUG oslo_concurrency.lockutils [None req-379abbe1-61a6-4c63-9167-2135a9f50e91 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Lock "4de350e2-55cc-4795-85e6-e58b5b362459-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1155.408585] env[63531]: INFO nova.compute.manager [None req-379abbe1-61a6-4c63-9167-2135a9f50e91 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] [instance: 4de350e2-55cc-4795-85e6-e58b5b362459] Terminating instance [ 1155.410379] env[63531]: DEBUG nova.compute.manager [None req-379abbe1-61a6-4c63-9167-2135a9f50e91 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] [instance: 4de350e2-55cc-4795-85e6-e58b5b362459] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1155.410577] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-379abbe1-61a6-4c63-9167-2135a9f50e91 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] [instance: 4de350e2-55cc-4795-85e6-e58b5b362459] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1155.411422] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-238f31e0-15c4-40f4-a244-27267f29f70a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.420190] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-379abbe1-61a6-4c63-9167-2135a9f50e91 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] [instance: 4de350e2-55cc-4795-85e6-e58b5b362459] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1155.420190] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-56bbc221-3fb4-4e9e-aa66-2921040b0852 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.425873] env[63531]: DEBUG oslo_vmware.api [None req-379abbe1-61a6-4c63-9167-2135a9f50e91 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Waiting for the task: (returnval){ [ 1155.425873] env[63531]: value = "task-1118737" [ 1155.425873] env[63531]: _type = "Task" [ 1155.425873] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.433337] env[63531]: DEBUG oslo_vmware.api [None req-379abbe1-61a6-4c63-9167-2135a9f50e91 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Task: {'id': task-1118737, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.437499] env[63531]: DEBUG nova.compute.manager [req-d6d3a0ef-d3c2-4dc5-ad8b-0b149dfe4762 req-3a776b07-3913-481c-989e-25c5a8f86b28 service nova] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Received event network-changed-1d65a3d9-a09b-4266-8a05-6abd7ddd18c7 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1155.437685] env[63531]: DEBUG nova.compute.manager [req-d6d3a0ef-d3c2-4dc5-ad8b-0b149dfe4762 req-3a776b07-3913-481c-989e-25c5a8f86b28 service nova] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Refreshing instance network info cache due to event network-changed-1d65a3d9-a09b-4266-8a05-6abd7ddd18c7. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1155.437908] env[63531]: DEBUG oslo_concurrency.lockutils [req-d6d3a0ef-d3c2-4dc5-ad8b-0b149dfe4762 req-3a776b07-3913-481c-989e-25c5a8f86b28 service nova] Acquiring lock "refresh_cache-98853767-791e-4568-8bb8-30dfcc473ee3" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1155.438093] env[63531]: DEBUG oslo_concurrency.lockutils [req-d6d3a0ef-d3c2-4dc5-ad8b-0b149dfe4762 req-3a776b07-3913-481c-989e-25c5a8f86b28 service nova] Acquired lock "refresh_cache-98853767-791e-4568-8bb8-30dfcc473ee3" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1155.438284] env[63531]: DEBUG nova.network.neutron [req-d6d3a0ef-d3c2-4dc5-ad8b-0b149dfe4762 req-3a776b07-3913-481c-989e-25c5a8f86b28 service nova] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Refreshing network info cache for port 1d65a3d9-a09b-4266-8a05-6abd7ddd18c7 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1155.562023] env[63531]: DEBUG oslo_vmware.api [None req-3447c1a3-dd26-481f-97eb-b24f05a0cc18 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118736, 'name': PowerOffVM_Task, 'duration_secs': 0.186792} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.562276] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-3447c1a3-dd26-481f-97eb-b24f05a0cc18 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1155.562458] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-3447c1a3-dd26-481f-97eb-b24f05a0cc18 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1155.562710] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5aecb400-825c-4506-9cad-504c898dccda {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.630115] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-3447c1a3-dd26-481f-97eb-b24f05a0cc18 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1155.630366] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-3447c1a3-dd26-481f-97eb-b24f05a0cc18 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Deleting contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1155.630674] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-3447c1a3-dd26-481f-97eb-b24f05a0cc18 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Deleting the datastore file [datastore2] 225e11e5-0558-4ad5-acd4-2a69888fe2d1 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1155.631027] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d2c16298-694f-451d-a209-06e6ca793b8a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.637953] env[63531]: DEBUG oslo_vmware.api [None req-3447c1a3-dd26-481f-97eb-b24f05a0cc18 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for the task: (returnval){ [ 1155.637953] env[63531]: value = "task-1118739" [ 1155.637953] env[63531]: _type = "Task" [ 1155.637953] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.645529] env[63531]: DEBUG oslo_vmware.api [None req-3447c1a3-dd26-481f-97eb-b24f05a0cc18 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118739, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.936485] env[63531]: DEBUG oslo_vmware.api [None req-379abbe1-61a6-4c63-9167-2135a9f50e91 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Task: {'id': task-1118737, 'name': PowerOffVM_Task, 'duration_secs': 0.209963} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.936764] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-379abbe1-61a6-4c63-9167-2135a9f50e91 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] [instance: 4de350e2-55cc-4795-85e6-e58b5b362459] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1155.936764] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-379abbe1-61a6-4c63-9167-2135a9f50e91 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] [instance: 4de350e2-55cc-4795-85e6-e58b5b362459] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1155.937058] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f25ed434-7b01-4666-80cd-c570fcc1d1cf {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.002461] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-379abbe1-61a6-4c63-9167-2135a9f50e91 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] [instance: 4de350e2-55cc-4795-85e6-e58b5b362459] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1156.002701] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-379abbe1-61a6-4c63-9167-2135a9f50e91 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] [instance: 4de350e2-55cc-4795-85e6-e58b5b362459] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1156.002879] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-379abbe1-61a6-4c63-9167-2135a9f50e91 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Deleting the datastore file [datastore1] 4de350e2-55cc-4795-85e6-e58b5b362459 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1156.005220] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1e7708b0-d45f-4eea-857f-cf9cc114207a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.012473] env[63531]: DEBUG oslo_vmware.api [None req-379abbe1-61a6-4c63-9167-2135a9f50e91 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Waiting for the task: (returnval){ [ 1156.012473] env[63531]: value = "task-1118741" [ 1156.012473] env[63531]: _type = "Task" [ 1156.012473] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.020253] env[63531]: DEBUG oslo_vmware.api [None req-379abbe1-61a6-4c63-9167-2135a9f50e91 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Task: {'id': task-1118741, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.147656] env[63531]: DEBUG oslo_vmware.api [None req-3447c1a3-dd26-481f-97eb-b24f05a0cc18 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Task: {'id': task-1118739, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145994} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.147917] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-3447c1a3-dd26-481f-97eb-b24f05a0cc18 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1156.148115] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-3447c1a3-dd26-481f-97eb-b24f05a0cc18 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Deleted contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1156.148306] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-3447c1a3-dd26-481f-97eb-b24f05a0cc18 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1156.148549] env[63531]: INFO nova.compute.manager [None req-3447c1a3-dd26-481f-97eb-b24f05a0cc18 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1156.148816] env[63531]: DEBUG oslo.service.loopingcall [None req-3447c1a3-dd26-481f-97eb-b24f05a0cc18 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1156.149038] env[63531]: DEBUG nova.compute.manager [-] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1156.149136] env[63531]: DEBUG nova.network.neutron [-] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1156.151331] env[63531]: DEBUG nova.network.neutron [req-d6d3a0ef-d3c2-4dc5-ad8b-0b149dfe4762 req-3a776b07-3913-481c-989e-25c5a8f86b28 service nova] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Updated VIF entry in instance network info cache for port 1d65a3d9-a09b-4266-8a05-6abd7ddd18c7. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1156.151656] env[63531]: DEBUG nova.network.neutron [req-d6d3a0ef-d3c2-4dc5-ad8b-0b149dfe4762 req-3a776b07-3913-481c-989e-25c5a8f86b28 service nova] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Updating instance_info_cache with network_info: [{"id": "1d65a3d9-a09b-4266-8a05-6abd7ddd18c7", "address": "fa:16:3e:14:b8:48", "network": {"id": "caf86428-59a5-4e50-9135-2ab8473af64d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1673994696-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ba8b284854242c392aec5326e996239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d65a3d9-a0", "ovs_interfaceid": "1d65a3d9-a09b-4266-8a05-6abd7ddd18c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1156.523215] env[63531]: DEBUG oslo_vmware.api [None req-379abbe1-61a6-4c63-9167-2135a9f50e91 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Task: {'id': task-1118741, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.179171} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.523484] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-379abbe1-61a6-4c63-9167-2135a9f50e91 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1156.523672] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-379abbe1-61a6-4c63-9167-2135a9f50e91 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] [instance: 4de350e2-55cc-4795-85e6-e58b5b362459] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1156.523860] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-379abbe1-61a6-4c63-9167-2135a9f50e91 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] [instance: 4de350e2-55cc-4795-85e6-e58b5b362459] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1156.524047] env[63531]: INFO nova.compute.manager [None req-379abbe1-61a6-4c63-9167-2135a9f50e91 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] [instance: 4de350e2-55cc-4795-85e6-e58b5b362459] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1156.524295] env[63531]: DEBUG oslo.service.loopingcall [None req-379abbe1-61a6-4c63-9167-2135a9f50e91 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1156.524488] env[63531]: DEBUG nova.compute.manager [-] [instance: 4de350e2-55cc-4795-85e6-e58b5b362459] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1156.524584] env[63531]: DEBUG nova.network.neutron [-] [instance: 4de350e2-55cc-4795-85e6-e58b5b362459] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1156.533437] env[63531]: DEBUG nova.compute.manager [req-99eef287-9db2-4bd5-aadb-cf43ad117189 req-f2b5b18a-3cab-4d1c-b68d-d8ac9f6d8187 service nova] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Received event network-vif-plugged-baf61e1c-97ff-411c-a0f7-415896653c60 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1156.533701] env[63531]: DEBUG oslo_concurrency.lockutils [req-99eef287-9db2-4bd5-aadb-cf43ad117189 req-f2b5b18a-3cab-4d1c-b68d-d8ac9f6d8187 service nova] Acquiring lock "98853767-791e-4568-8bb8-30dfcc473ee3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1156.533947] env[63531]: DEBUG oslo_concurrency.lockutils [req-99eef287-9db2-4bd5-aadb-cf43ad117189 req-f2b5b18a-3cab-4d1c-b68d-d8ac9f6d8187 service nova] Lock "98853767-791e-4568-8bb8-30dfcc473ee3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1156.534317] env[63531]: DEBUG oslo_concurrency.lockutils [req-99eef287-9db2-4bd5-aadb-cf43ad117189 req-f2b5b18a-3cab-4d1c-b68d-d8ac9f6d8187 service nova] Lock "98853767-791e-4568-8bb8-30dfcc473ee3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1156.534407] env[63531]: DEBUG nova.compute.manager [req-99eef287-9db2-4bd5-aadb-cf43ad117189 req-f2b5b18a-3cab-4d1c-b68d-d8ac9f6d8187 service nova] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] No waiting events found dispatching network-vif-plugged-baf61e1c-97ff-411c-a0f7-415896653c60 {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1156.534562] env[63531]: WARNING nova.compute.manager [req-99eef287-9db2-4bd5-aadb-cf43ad117189 req-f2b5b18a-3cab-4d1c-b68d-d8ac9f6d8187 service nova] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Received unexpected event network-vif-plugged-baf61e1c-97ff-411c-a0f7-415896653c60 for instance with vm_state active and task_state None. [ 1156.627734] env[63531]: DEBUG nova.network.neutron [None req-5eadccc3-df04-41f8-b336-4b32353995e2 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Successfully updated port: baf61e1c-97ff-411c-a0f7-415896653c60 {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1156.654086] env[63531]: DEBUG oslo_concurrency.lockutils [req-d6d3a0ef-d3c2-4dc5-ad8b-0b149dfe4762 req-3a776b07-3913-481c-989e-25c5a8f86b28 service nova] Releasing lock "refresh_cache-98853767-791e-4568-8bb8-30dfcc473ee3" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1156.875415] env[63531]: DEBUG nova.network.neutron [-] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1157.131954] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5eadccc3-df04-41f8-b336-4b32353995e2 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "refresh_cache-98853767-791e-4568-8bb8-30dfcc473ee3" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1157.132280] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5eadccc3-df04-41f8-b336-4b32353995e2 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquired lock "refresh_cache-98853767-791e-4568-8bb8-30dfcc473ee3" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1157.132523] env[63531]: DEBUG nova.network.neutron [None req-5eadccc3-df04-41f8-b336-4b32353995e2 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1157.241181] env[63531]: DEBUG nova.network.neutron [-] [instance: 4de350e2-55cc-4795-85e6-e58b5b362459] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1157.378450] env[63531]: INFO nova.compute.manager [-] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Took 1.23 seconds to deallocate network for instance. [ 1157.466289] env[63531]: DEBUG nova.compute.manager [req-273090ae-73e0-48a1-a489-2f2a274917f2 req-b303f9d5-67b1-413f-b89e-e8dbcf109daa service nova] [instance: 225e11e5-0558-4ad5-acd4-2a69888fe2d1] Received event network-vif-deleted-8b6b5f30-0bda-4851-9288-3c3a856e8a6c {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1157.466460] env[63531]: DEBUG nova.compute.manager [req-273090ae-73e0-48a1-a489-2f2a274917f2 req-b303f9d5-67b1-413f-b89e-e8dbcf109daa service nova] [instance: 4de350e2-55cc-4795-85e6-e58b5b362459] Received event network-vif-deleted-85e94a7c-07ba-4f11-b4f0-ac8f6ef24c8e {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1157.667629] env[63531]: WARNING nova.network.neutron [None req-5eadccc3-df04-41f8-b336-4b32353995e2 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] caf86428-59a5-4e50-9135-2ab8473af64d already exists in list: networks containing: ['caf86428-59a5-4e50-9135-2ab8473af64d']. ignoring it [ 1157.744016] env[63531]: INFO nova.compute.manager [-] [instance: 4de350e2-55cc-4795-85e6-e58b5b362459] Took 1.22 seconds to deallocate network for instance. [ 1157.885791] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3447c1a3-dd26-481f-97eb-b24f05a0cc18 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1157.886086] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3447c1a3-dd26-481f-97eb-b24f05a0cc18 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1157.886314] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3447c1a3-dd26-481f-97eb-b24f05a0cc18 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1157.913427] env[63531]: INFO nova.scheduler.client.report [None req-3447c1a3-dd26-481f-97eb-b24f05a0cc18 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Deleted allocations for instance 225e11e5-0558-4ad5-acd4-2a69888fe2d1 [ 1157.941771] env[63531]: DEBUG nova.network.neutron [None req-5eadccc3-df04-41f8-b336-4b32353995e2 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Updating instance_info_cache with network_info: [{"id": "1d65a3d9-a09b-4266-8a05-6abd7ddd18c7", "address": "fa:16:3e:14:b8:48", "network": {"id": "caf86428-59a5-4e50-9135-2ab8473af64d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1673994696-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ba8b284854242c392aec5326e996239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d65a3d9-a0", "ovs_interfaceid": "1d65a3d9-a09b-4266-8a05-6abd7ddd18c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "baf61e1c-97ff-411c-a0f7-415896653c60", "address": "fa:16:3e:88:db:88", "network": {"id": "caf86428-59a5-4e50-9135-2ab8473af64d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1673994696-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ba8b284854242c392aec5326e996239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbaf61e1c-97", "ovs_interfaceid": "baf61e1c-97ff-411c-a0f7-415896653c60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1158.252364] env[63531]: DEBUG oslo_concurrency.lockutils [None req-379abbe1-61a6-4c63-9167-2135a9f50e91 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1158.252675] env[63531]: DEBUG oslo_concurrency.lockutils [None req-379abbe1-61a6-4c63-9167-2135a9f50e91 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1158.252871] env[63531]: DEBUG nova.objects.instance [None req-379abbe1-61a6-4c63-9167-2135a9f50e91 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Lazy-loading 'resources' on Instance uuid 4de350e2-55cc-4795-85e6-e58b5b362459 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1158.421918] env[63531]: DEBUG oslo_concurrency.lockutils [None req-3447c1a3-dd26-481f-97eb-b24f05a0cc18 tempest-DeleteServersTestJSON-1328056822 tempest-DeleteServersTestJSON-1328056822-project-member] Lock "225e11e5-0558-4ad5-acd4-2a69888fe2d1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.395s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1158.444760] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5eadccc3-df04-41f8-b336-4b32353995e2 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Releasing lock "refresh_cache-98853767-791e-4568-8bb8-30dfcc473ee3" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1158.445426] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5eadccc3-df04-41f8-b336-4b32353995e2 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "98853767-791e-4568-8bb8-30dfcc473ee3" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1158.445594] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5eadccc3-df04-41f8-b336-4b32353995e2 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquired lock "98853767-791e-4568-8bb8-30dfcc473ee3" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1158.446444] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-063521c6-b568-4f9c-9f2b-6803dac96379 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.464072] env[63531]: DEBUG nova.virt.hardware [None req-5eadccc3-df04-41f8-b336-4b32353995e2 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1158.464286] env[63531]: DEBUG nova.virt.hardware [None req-5eadccc3-df04-41f8-b336-4b32353995e2 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1158.464450] env[63531]: DEBUG nova.virt.hardware [None req-5eadccc3-df04-41f8-b336-4b32353995e2 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1158.464637] env[63531]: DEBUG nova.virt.hardware [None req-5eadccc3-df04-41f8-b336-4b32353995e2 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1158.464790] env[63531]: DEBUG nova.virt.hardware [None req-5eadccc3-df04-41f8-b336-4b32353995e2 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1158.465039] env[63531]: DEBUG nova.virt.hardware [None req-5eadccc3-df04-41f8-b336-4b32353995e2 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1158.465198] env[63531]: DEBUG nova.virt.hardware [None req-5eadccc3-df04-41f8-b336-4b32353995e2 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1158.465404] env[63531]: DEBUG nova.virt.hardware [None req-5eadccc3-df04-41f8-b336-4b32353995e2 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1158.465572] env[63531]: DEBUG nova.virt.hardware [None req-5eadccc3-df04-41f8-b336-4b32353995e2 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1158.465744] env[63531]: DEBUG nova.virt.hardware [None req-5eadccc3-df04-41f8-b336-4b32353995e2 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1158.465926] env[63531]: DEBUG nova.virt.hardware [None req-5eadccc3-df04-41f8-b336-4b32353995e2 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1158.472168] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5eadccc3-df04-41f8-b336-4b32353995e2 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Reconfiguring VM to attach interface {{(pid=63531) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 1158.472498] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0a3d253b-36fe-4d58-aafc-10fbbe6bec9f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.489809] env[63531]: DEBUG oslo_vmware.api [None req-5eadccc3-df04-41f8-b336-4b32353995e2 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for the task: (returnval){ [ 1158.489809] env[63531]: value = "task-1118743" [ 1158.489809] env[63531]: _type = "Task" [ 1158.489809] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.497148] env[63531]: DEBUG oslo_vmware.api [None req-5eadccc3-df04-41f8-b336-4b32353995e2 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118743, 'name': ReconfigVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.559662] env[63531]: DEBUG nova.compute.manager [req-d15f9841-8e6d-45b6-8f3d-7b38aecf3115 req-a608c625-5125-44da-a34d-3785a85b145e service nova] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Received event network-changed-baf61e1c-97ff-411c-a0f7-415896653c60 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1158.559830] env[63531]: DEBUG nova.compute.manager [req-d15f9841-8e6d-45b6-8f3d-7b38aecf3115 req-a608c625-5125-44da-a34d-3785a85b145e service nova] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Refreshing instance network info cache due to event network-changed-baf61e1c-97ff-411c-a0f7-415896653c60. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1158.560074] env[63531]: DEBUG oslo_concurrency.lockutils [req-d15f9841-8e6d-45b6-8f3d-7b38aecf3115 req-a608c625-5125-44da-a34d-3785a85b145e service nova] Acquiring lock "refresh_cache-98853767-791e-4568-8bb8-30dfcc473ee3" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1158.560227] env[63531]: DEBUG oslo_concurrency.lockutils [req-d15f9841-8e6d-45b6-8f3d-7b38aecf3115 req-a608c625-5125-44da-a34d-3785a85b145e service nova] Acquired lock "refresh_cache-98853767-791e-4568-8bb8-30dfcc473ee3" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1158.560395] env[63531]: DEBUG nova.network.neutron [req-d15f9841-8e6d-45b6-8f3d-7b38aecf3115 req-a608c625-5125-44da-a34d-3785a85b145e service nova] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Refreshing network info cache for port baf61e1c-97ff-411c-a0f7-415896653c60 {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1158.825799] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f88cbb8-4d9f-43b3-bcda-939eaae8fd33 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.834188] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be66147f-a1f3-4bd0-ac31-46e8c5bec07b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.865167] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af0c28e7-e910-4881-8c0b-642dede2a1fe {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.873121] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdec97da-9562-48ce-ba74-4ead6955c35d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.886942] env[63531]: DEBUG nova.compute.provider_tree [None req-379abbe1-61a6-4c63-9167-2135a9f50e91 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1159.002141] env[63531]: DEBUG oslo_vmware.api [None req-5eadccc3-df04-41f8-b336-4b32353995e2 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118743, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.276462] env[63531]: DEBUG nova.network.neutron [req-d15f9841-8e6d-45b6-8f3d-7b38aecf3115 req-a608c625-5125-44da-a34d-3785a85b145e service nova] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Updated VIF entry in instance network info cache for port baf61e1c-97ff-411c-a0f7-415896653c60. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1159.276904] env[63531]: DEBUG nova.network.neutron [req-d15f9841-8e6d-45b6-8f3d-7b38aecf3115 req-a608c625-5125-44da-a34d-3785a85b145e service nova] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Updating instance_info_cache with network_info: [{"id": "1d65a3d9-a09b-4266-8a05-6abd7ddd18c7", "address": "fa:16:3e:14:b8:48", "network": {"id": "caf86428-59a5-4e50-9135-2ab8473af64d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1673994696-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ba8b284854242c392aec5326e996239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d65a3d9-a0", "ovs_interfaceid": "1d65a3d9-a09b-4266-8a05-6abd7ddd18c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "baf61e1c-97ff-411c-a0f7-415896653c60", "address": "fa:16:3e:88:db:88", "network": {"id": "caf86428-59a5-4e50-9135-2ab8473af64d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1673994696-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ba8b284854242c392aec5326e996239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbaf61e1c-97", "ovs_interfaceid": "baf61e1c-97ff-411c-a0f7-415896653c60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1159.389963] env[63531]: DEBUG nova.scheduler.client.report [None req-379abbe1-61a6-4c63-9167-2135a9f50e91 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1159.500737] env[63531]: DEBUG oslo_vmware.api [None req-5eadccc3-df04-41f8-b336-4b32353995e2 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118743, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.780248] env[63531]: DEBUG oslo_concurrency.lockutils [req-d15f9841-8e6d-45b6-8f3d-7b38aecf3115 req-a608c625-5125-44da-a34d-3785a85b145e service nova] Releasing lock "refresh_cache-98853767-791e-4568-8bb8-30dfcc473ee3" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1159.895578] env[63531]: DEBUG oslo_concurrency.lockutils [None req-379abbe1-61a6-4c63-9167-2135a9f50e91 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.643s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1159.913334] env[63531]: INFO nova.scheduler.client.report [None req-379abbe1-61a6-4c63-9167-2135a9f50e91 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Deleted allocations for instance 4de350e2-55cc-4795-85e6-e58b5b362459 [ 1160.002304] env[63531]: DEBUG oslo_vmware.api [None req-5eadccc3-df04-41f8-b336-4b32353995e2 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118743, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.421729] env[63531]: DEBUG oslo_concurrency.lockutils [None req-379abbe1-61a6-4c63-9167-2135a9f50e91 tempest-ServerGroupTestJSON-1500654820 tempest-ServerGroupTestJSON-1500654820-project-member] Lock "4de350e2-55cc-4795-85e6-e58b5b362459" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.016s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1160.502197] env[63531]: DEBUG oslo_vmware.api [None req-5eadccc3-df04-41f8-b336-4b32353995e2 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118743, 'name': ReconfigVM_Task, 'duration_secs': 1.600241} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.502763] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5eadccc3-df04-41f8-b336-4b32353995e2 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Releasing lock "98853767-791e-4568-8bb8-30dfcc473ee3" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1160.503029] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-5eadccc3-df04-41f8-b336-4b32353995e2 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Reconfigured VM to attach interface {{(pid=63531) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 1161.007686] env[63531]: DEBUG oslo_concurrency.lockutils [None req-5eadccc3-df04-41f8-b336-4b32353995e2 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "interface-98853767-791e-4568-8bb8-30dfcc473ee3-baf61e1c-97ff-411c-a0f7-415896653c60" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.027s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1162.308740] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1162.693830] env[63531]: DEBUG oslo_concurrency.lockutils [None req-94e00d23-9845-4c2b-9134-9ea961c3629f tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "interface-98853767-791e-4568-8bb8-30dfcc473ee3-baf61e1c-97ff-411c-a0f7-415896653c60" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1162.694109] env[63531]: DEBUG oslo_concurrency.lockutils [None req-94e00d23-9845-4c2b-9134-9ea961c3629f tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "interface-98853767-791e-4568-8bb8-30dfcc473ee3-baf61e1c-97ff-411c-a0f7-415896653c60" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1162.777326] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1162.778371] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Starting heal instance info cache {{(pid=63531) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1163.197631] env[63531]: DEBUG oslo_concurrency.lockutils [None req-94e00d23-9845-4c2b-9134-9ea961c3629f tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "98853767-791e-4568-8bb8-30dfcc473ee3" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1163.197823] env[63531]: DEBUG oslo_concurrency.lockutils [None req-94e00d23-9845-4c2b-9134-9ea961c3629f tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquired lock "98853767-791e-4568-8bb8-30dfcc473ee3" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1163.198777] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17ef3a9c-b329-41cb-8912-3315b3f301c6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.218209] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-576a93e9-bd57-46a5-b413-1ba14dda1987 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.244423] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-94e00d23-9845-4c2b-9134-9ea961c3629f tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Reconfiguring VM to detach interface {{(pid=63531) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 1163.244752] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c55b56ac-4daa-4811-adef-ff9416e152b4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.263406] env[63531]: DEBUG oslo_vmware.api [None req-94e00d23-9845-4c2b-9134-9ea961c3629f tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for the task: (returnval){ [ 1163.263406] env[63531]: value = "task-1118744" [ 1163.263406] env[63531]: _type = "Task" [ 1163.263406] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.272972] env[63531]: DEBUG oslo_vmware.api [None req-94e00d23-9845-4c2b-9134-9ea961c3629f tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118744, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.773595] env[63531]: DEBUG oslo_vmware.api [None req-94e00d23-9845-4c2b-9134-9ea961c3629f tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118744, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.275829] env[63531]: DEBUG oslo_vmware.api [None req-94e00d23-9845-4c2b-9134-9ea961c3629f tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118744, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.775350] env[63531]: DEBUG oslo_vmware.api [None req-94e00d23-9845-4c2b-9134-9ea961c3629f tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118744, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.275882] env[63531]: DEBUG oslo_vmware.api [None req-94e00d23-9845-4c2b-9134-9ea961c3629f tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118744, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.777415] env[63531]: DEBUG oslo_vmware.api [None req-94e00d23-9845-4c2b-9134-9ea961c3629f tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118744, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.792075] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Didn't find any instances for network info cache update. {{(pid=63531) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1165.792234] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1165.792452] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1165.792632] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1165.792797] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63531) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1166.278435] env[63531]: DEBUG oslo_vmware.api [None req-94e00d23-9845-4c2b-9134-9ea961c3629f tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118744, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.778624] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1166.778972] env[63531]: DEBUG oslo_vmware.api [None req-94e00d23-9845-4c2b-9134-9ea961c3629f tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118744, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.779168] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1166.779333] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager.update_available_resource {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1167.279044] env[63531]: DEBUG oslo_vmware.api [None req-94e00d23-9845-4c2b-9134-9ea961c3629f tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118744, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.282037] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1167.282266] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1167.282431] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1167.282583] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63531) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1167.283327] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e76f128f-5a18-4e04-95d8-48ff495cdd60 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.290304] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c153d979-018c-4109-a936-019aa0c29404 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.303237] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e409cc7-5b91-4872-8a46-1a2498c72468 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.309326] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f8d3232-29b8-4990-a4f8-437b486f88c4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.337927] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180718MB free_disk=170GB free_vcpus=48 pci_devices=None {{(pid=63531) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1167.338065] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1167.338245] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1167.779510] env[63531]: DEBUG oslo_vmware.api [None req-94e00d23-9845-4c2b-9134-9ea961c3629f tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118744, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.280425] env[63531]: DEBUG oslo_vmware.api [None req-94e00d23-9845-4c2b-9134-9ea961c3629f tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118744, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.363031] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance ccc8f366-5849-45a6-9745-ff9b2be9c3bb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1168.363197] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 98853767-791e-4568-8bb8-30dfcc473ee3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1168.363334] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1168.363453] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Instance 2d51997d-f88c-468c-9a56-0331a06ca56c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63531) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1168.363627] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=63531) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1168.363767] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=63531) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1168.420968] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4687425c-57a0-4db3-9f73-1ffb2ce2b90d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.428197] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc9c293c-6851-4d79-a785-5fb9c3f0f9ee {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.456557] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6c0eaa8-f32e-40b6-bdf0-32510bc43020 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.463445] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b6bd199-1dbb-42b6-83c2-bf40c2fc8a7f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.478609] env[63531]: DEBUG nova.compute.provider_tree [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1168.780867] env[63531]: DEBUG oslo_vmware.api [None req-94e00d23-9845-4c2b-9134-9ea961c3629f tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118744, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.981532] env[63531]: DEBUG nova.scheduler.client.report [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1169.280288] env[63531]: DEBUG oslo_vmware.api [None req-94e00d23-9845-4c2b-9134-9ea961c3629f tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118744, 'name': ReconfigVM_Task, 'duration_secs': 5.766265} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1169.280550] env[63531]: DEBUG oslo_concurrency.lockutils [None req-94e00d23-9845-4c2b-9134-9ea961c3629f tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Releasing lock "98853767-791e-4568-8bb8-30dfcc473ee3" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1169.280765] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-94e00d23-9845-4c2b-9134-9ea961c3629f tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Reconfigured VM to detach interface {{(pid=63531) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 1169.486377] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63531) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1169.486569] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.148s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1169.843533] env[63531]: DEBUG oslo_concurrency.lockutils [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquiring lock "3794f44c-8f6d-4bab-b9e0-9f6c819e1a97" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1169.843887] env[63531]: DEBUG oslo_concurrency.lockutils [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "3794f44c-8f6d-4bab-b9e0-9f6c819e1a97" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1169.843979] env[63531]: INFO nova.compute.manager [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Shelving [ 1170.184097] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f2825085-54b5-490c-baee-b80e728272a0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "98853767-791e-4568-8bb8-30dfcc473ee3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1170.184375] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f2825085-54b5-490c-baee-b80e728272a0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "98853767-791e-4568-8bb8-30dfcc473ee3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1170.184592] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f2825085-54b5-490c-baee-b80e728272a0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "98853767-791e-4568-8bb8-30dfcc473ee3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1170.184780] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f2825085-54b5-490c-baee-b80e728272a0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "98853767-791e-4568-8bb8-30dfcc473ee3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1170.184957] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f2825085-54b5-490c-baee-b80e728272a0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "98853767-791e-4568-8bb8-30dfcc473ee3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1170.187419] env[63531]: INFO nova.compute.manager [None req-f2825085-54b5-490c-baee-b80e728272a0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Terminating instance [ 1170.189117] env[63531]: DEBUG nova.compute.manager [None req-f2825085-54b5-490c-baee-b80e728272a0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1170.189317] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f2825085-54b5-490c-baee-b80e728272a0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1170.190167] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6de447fc-faa9-464f-b7c0-137ff3367843 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.197689] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2825085-54b5-490c-baee-b80e728272a0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1170.197916] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1acab72e-7f30-4eb5-ad22-70114594df1d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.203931] env[63531]: DEBUG oslo_vmware.api [None req-f2825085-54b5-490c-baee-b80e728272a0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for the task: (returnval){ [ 1170.203931] env[63531]: value = "task-1118745" [ 1170.203931] env[63531]: _type = "Task" [ 1170.203931] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.211500] env[63531]: DEBUG oslo_vmware.api [None req-f2825085-54b5-490c-baee-b80e728272a0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118745, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.351548] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1170.351811] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-775e7ddd-4599-4378-8a3d-c6aa82547f6d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.359014] env[63531]: DEBUG oslo_vmware.api [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 1170.359014] env[63531]: value = "task-1118746" [ 1170.359014] env[63531]: _type = "Task" [ 1170.359014] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.366979] env[63531]: DEBUG oslo_vmware.api [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118746, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.485166] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1170.573299] env[63531]: DEBUG oslo_concurrency.lockutils [None req-94e00d23-9845-4c2b-9134-9ea961c3629f tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "refresh_cache-98853767-791e-4568-8bb8-30dfcc473ee3" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1170.573533] env[63531]: DEBUG oslo_concurrency.lockutils [None req-94e00d23-9845-4c2b-9134-9ea961c3629f tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquired lock "refresh_cache-98853767-791e-4568-8bb8-30dfcc473ee3" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1170.573754] env[63531]: DEBUG nova.network.neutron [None req-94e00d23-9845-4c2b-9134-9ea961c3629f tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1170.715559] env[63531]: DEBUG oslo_vmware.api [None req-f2825085-54b5-490c-baee-b80e728272a0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118745, 'name': PowerOffVM_Task, 'duration_secs': 0.213701} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.715812] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2825085-54b5-490c-baee-b80e728272a0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1170.715983] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f2825085-54b5-490c-baee-b80e728272a0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1170.716277] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a74b8d70-31db-4d77-9bf1-103cb799a41f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.783041] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f2825085-54b5-490c-baee-b80e728272a0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1170.783209] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f2825085-54b5-490c-baee-b80e728272a0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1170.783396] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2825085-54b5-490c-baee-b80e728272a0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Deleting the datastore file [datastore1] 98853767-791e-4568-8bb8-30dfcc473ee3 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1170.783642] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-788d2abe-51e5-4fab-b3a2-65fb908e1927 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.789614] env[63531]: DEBUG oslo_vmware.api [None req-f2825085-54b5-490c-baee-b80e728272a0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for the task: (returnval){ [ 1170.789614] env[63531]: value = "task-1118748" [ 1170.789614] env[63531]: _type = "Task" [ 1170.789614] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.797353] env[63531]: DEBUG oslo_vmware.api [None req-f2825085-54b5-490c-baee-b80e728272a0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118748, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.869034] env[63531]: DEBUG oslo_vmware.api [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118746, 'name': PowerOffVM_Task, 'duration_secs': 0.181597} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.869384] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1170.870068] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dfaeca5-fca8-4bba-a91a-9bf21aadf28a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.888014] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74fa140b-83ce-49b8-b084-b705d98d17b1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.286448] env[63531]: INFO nova.network.neutron [None req-94e00d23-9845-4c2b-9134-9ea961c3629f tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Port baf61e1c-97ff-411c-a0f7-415896653c60 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1171.286823] env[63531]: DEBUG nova.network.neutron [None req-94e00d23-9845-4c2b-9134-9ea961c3629f tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Updating instance_info_cache with network_info: [{"id": "1d65a3d9-a09b-4266-8a05-6abd7ddd18c7", "address": "fa:16:3e:14:b8:48", "network": {"id": "caf86428-59a5-4e50-9135-2ab8473af64d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1673994696-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ba8b284854242c392aec5326e996239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d65a3d9-a0", "ovs_interfaceid": "1d65a3d9-a09b-4266-8a05-6abd7ddd18c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1171.299746] env[63531]: DEBUG oslo_vmware.api [None req-f2825085-54b5-490c-baee-b80e728272a0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118748, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.150178} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1171.300602] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2825085-54b5-490c-baee-b80e728272a0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1171.300794] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f2825085-54b5-490c-baee-b80e728272a0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1171.301019] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-f2825085-54b5-490c-baee-b80e728272a0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1171.301205] env[63531]: INFO nova.compute.manager [None req-f2825085-54b5-490c-baee-b80e728272a0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1171.301443] env[63531]: DEBUG oslo.service.loopingcall [None req-f2825085-54b5-490c-baee-b80e728272a0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1171.301631] env[63531]: DEBUG nova.compute.manager [-] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1171.301726] env[63531]: DEBUG nova.network.neutron [-] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1171.397840] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Creating Snapshot of the VM instance {{(pid=63531) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1171.398167] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-eef6691f-292a-4a7b-aa18-65f418d6fce9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.405890] env[63531]: DEBUG oslo_vmware.api [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 1171.405890] env[63531]: value = "task-1118749" [ 1171.405890] env[63531]: _type = "Task" [ 1171.405890] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1171.413930] env[63531]: DEBUG oslo_vmware.api [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118749, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.789469] env[63531]: DEBUG oslo_concurrency.lockutils [None req-94e00d23-9845-4c2b-9134-9ea961c3629f tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Releasing lock "refresh_cache-98853767-791e-4568-8bb8-30dfcc473ee3" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1171.918728] env[63531]: DEBUG oslo_vmware.api [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118749, 'name': CreateSnapshot_Task, 'duration_secs': 0.478359} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1171.919138] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Created Snapshot of the VM instance {{(pid=63531) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1171.919906] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5d1851a-9907-417f-beba-0df21c815a00 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.007972] env[63531]: DEBUG nova.compute.manager [req-127481ab-6acc-4b67-975c-7915aec753ed req-4960bf84-9e9d-4535-825c-8f1b23f8577d service nova] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Received event network-vif-deleted-1d65a3d9-a09b-4266-8a05-6abd7ddd18c7 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1172.007972] env[63531]: INFO nova.compute.manager [req-127481ab-6acc-4b67-975c-7915aec753ed req-4960bf84-9e9d-4535-825c-8f1b23f8577d service nova] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Neutron deleted interface 1d65a3d9-a09b-4266-8a05-6abd7ddd18c7; detaching it from the instance and deleting it from the info cache [ 1172.008282] env[63531]: DEBUG nova.network.neutron [req-127481ab-6acc-4b67-975c-7915aec753ed req-4960bf84-9e9d-4535-825c-8f1b23f8577d service nova] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1172.295028] env[63531]: DEBUG oslo_concurrency.lockutils [None req-94e00d23-9845-4c2b-9134-9ea961c3629f tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "interface-98853767-791e-4568-8bb8-30dfcc473ee3-baf61e1c-97ff-411c-a0f7-415896653c60" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.600s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1172.415680] env[63531]: DEBUG nova.network.neutron [-] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1172.438073] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Creating linked-clone VM from snapshot {{(pid=63531) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1172.438682] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-632c1f4e-a78e-48b4-99cd-127a7b56c462 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.448053] env[63531]: DEBUG oslo_vmware.api [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 1172.448053] env[63531]: value = "task-1118750" [ 1172.448053] env[63531]: _type = "Task" [ 1172.448053] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1172.456102] env[63531]: DEBUG oslo_vmware.api [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118750, 'name': CloneVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.510993] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-82dff946-8c8d-46f2-8e21-9d1de51aed8d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.521408] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d676597-56ca-4f46-b7d3-6270b186e9a4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.545700] env[63531]: DEBUG nova.compute.manager [req-127481ab-6acc-4b67-975c-7915aec753ed req-4960bf84-9e9d-4535-825c-8f1b23f8577d service nova] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Detach interface failed, port_id=1d65a3d9-a09b-4266-8a05-6abd7ddd18c7, reason: Instance 98853767-791e-4568-8bb8-30dfcc473ee3 could not be found. {{(pid=63531) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1172.918464] env[63531]: INFO nova.compute.manager [-] [instance: 98853767-791e-4568-8bb8-30dfcc473ee3] Took 1.62 seconds to deallocate network for instance. [ 1172.959561] env[63531]: DEBUG oslo_vmware.api [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118750, 'name': CloneVM_Task} progress is 94%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.426304] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f6077cec-0f8d-4210-b306-e36b1bf3c2f5 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "2d51997d-f88c-468c-9a56-0331a06ca56c" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1173.426687] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f6077cec-0f8d-4210-b306-e36b1bf3c2f5 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "2d51997d-f88c-468c-9a56-0331a06ca56c" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1173.426939] env[63531]: DEBUG nova.compute.manager [None req-f6077cec-0f8d-4210-b306-e36b1bf3c2f5 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1173.427979] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f2825085-54b5-490c-baee-b80e728272a0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1173.428278] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f2825085-54b5-490c-baee-b80e728272a0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1173.428551] env[63531]: DEBUG nova.objects.instance [None req-f2825085-54b5-490c-baee-b80e728272a0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lazy-loading 'resources' on Instance uuid 98853767-791e-4568-8bb8-30dfcc473ee3 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1173.430272] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b82a8547-5cf7-4c55-af57-b6f4c78278a9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.436954] env[63531]: DEBUG nova.compute.manager [None req-f6077cec-0f8d-4210-b306-e36b1bf3c2f5 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63531) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1173.437627] env[63531]: DEBUG nova.objects.instance [None req-f6077cec-0f8d-4210-b306-e36b1bf3c2f5 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lazy-loading 'flavor' on Instance uuid 2d51997d-f88c-468c-9a56-0331a06ca56c {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1173.459011] env[63531]: DEBUG oslo_vmware.api [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118750, 'name': CloneVM_Task} progress is 95%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.942332] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6077cec-0f8d-4210-b306-e36b1bf3c2f5 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1173.942603] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0f255e7b-eef1-431d-a4dc-ee9afd80212c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.956483] env[63531]: DEBUG oslo_vmware.api [None req-f6077cec-0f8d-4210-b306-e36b1bf3c2f5 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 1173.956483] env[63531]: value = "task-1118751" [ 1173.956483] env[63531]: _type = "Task" [ 1173.956483] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1173.963046] env[63531]: DEBUG oslo_vmware.api [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118750, 'name': CloneVM_Task, 'duration_secs': 1.123602} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.963604] env[63531]: INFO nova.virt.vmwareapi.vmops [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Created linked-clone VM from snapshot [ 1173.964312] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4449e189-172f-4e0f-b450-f0ff76bef584 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.969509] env[63531]: DEBUG oslo_vmware.api [None req-f6077cec-0f8d-4210-b306-e36b1bf3c2f5 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118751, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.976899] env[63531]: DEBUG nova.virt.vmwareapi.images [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Uploading image c6bead00-8690-442e-a30b-eb4e618e7b4b {{(pid=63531) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1174.003189] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83ad33ed-9abb-488d-a36b-c369a2424ccd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.007608] env[63531]: DEBUG oslo_vmware.rw_handles [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1174.007608] env[63531]: value = "vm-244820" [ 1174.007608] env[63531]: _type = "VirtualMachine" [ 1174.007608] env[63531]: }. {{(pid=63531) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1174.007898] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-d4a04b50-1445-4ded-81e9-bf24f33f07b9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.014064] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a7f22f5-9f50-4d78-bf4a-25f610eb7965 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.017964] env[63531]: DEBUG oslo_vmware.rw_handles [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lease: (returnval){ [ 1174.017964] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]529d3a6f-6c85-186e-2cbe-f91da38e1fc1" [ 1174.017964] env[63531]: _type = "HttpNfcLease" [ 1174.017964] env[63531]: } obtained for exporting VM: (result){ [ 1174.017964] env[63531]: value = "vm-244820" [ 1174.017964] env[63531]: _type = "VirtualMachine" [ 1174.017964] env[63531]: }. {{(pid=63531) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1174.018290] env[63531]: DEBUG oslo_vmware.api [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the lease: (returnval){ [ 1174.018290] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]529d3a6f-6c85-186e-2cbe-f91da38e1fc1" [ 1174.018290] env[63531]: _type = "HttpNfcLease" [ 1174.018290] env[63531]: } to be ready. {{(pid=63531) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1174.046596] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-666787b8-3a68-419f-9fda-c4023face3ec {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.050428] env[63531]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1174.050428] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]529d3a6f-6c85-186e-2cbe-f91da38e1fc1" [ 1174.050428] env[63531]: _type = "HttpNfcLease" [ 1174.050428] env[63531]: } is ready. {{(pid=63531) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1174.051099] env[63531]: DEBUG oslo_vmware.rw_handles [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1174.051099] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]529d3a6f-6c85-186e-2cbe-f91da38e1fc1" [ 1174.051099] env[63531]: _type = "HttpNfcLease" [ 1174.051099] env[63531]: }. {{(pid=63531) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1174.051792] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed8185ba-b3ff-45dc-a81c-a22a67a492a5 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.057178] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-058f7839-bc6d-48b7-936f-e49953f0499a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.063414] env[63531]: DEBUG oslo_vmware.rw_handles [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b6948b-1438-2ad5-3304-dd9d39532d8d/disk-0.vmdk from lease info. {{(pid=63531) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1174.063592] env[63531]: DEBUG oslo_vmware.rw_handles [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b6948b-1438-2ad5-3304-dd9d39532d8d/disk-0.vmdk for reading. {{(pid=63531) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1174.127046] env[63531]: DEBUG nova.compute.provider_tree [None req-f2825085-54b5-490c-baee-b80e728272a0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1174.160766] env[63531]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-3bf28fcd-d10c-4d89-9271-70c7de631ee2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.468556] env[63531]: DEBUG oslo_vmware.api [None req-f6077cec-0f8d-4210-b306-e36b1bf3c2f5 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118751, 'name': PowerOffVM_Task, 'duration_secs': 0.197205} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1174.468899] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6077cec-0f8d-4210-b306-e36b1bf3c2f5 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1174.469172] env[63531]: DEBUG nova.compute.manager [None req-f6077cec-0f8d-4210-b306-e36b1bf3c2f5 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1174.470034] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7519e827-0520-4972-bbb8-c22fe6f73d9d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.631233] env[63531]: DEBUG nova.scheduler.client.report [None req-f2825085-54b5-490c-baee-b80e728272a0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1174.982552] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f6077cec-0f8d-4210-b306-e36b1bf3c2f5 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "2d51997d-f88c-468c-9a56-0331a06ca56c" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.556s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1175.135982] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f2825085-54b5-490c-baee-b80e728272a0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.707s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1175.155818] env[63531]: INFO nova.scheduler.client.report [None req-f2825085-54b5-490c-baee-b80e728272a0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Deleted allocations for instance 98853767-791e-4568-8bb8-30dfcc473ee3 [ 1175.665368] env[63531]: DEBUG oslo_concurrency.lockutils [None req-f2825085-54b5-490c-baee-b80e728272a0 tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "98853767-791e-4568-8bb8-30dfcc473ee3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.481s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1175.844636] env[63531]: DEBUG nova.objects.instance [None req-9955744b-2aea-43b6-b714-a0d13076ad57 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lazy-loading 'flavor' on Instance uuid 2d51997d-f88c-468c-9a56-0331a06ca56c {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1176.350682] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9955744b-2aea-43b6-b714-a0d13076ad57 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "refresh_cache-2d51997d-f88c-468c-9a56-0331a06ca56c" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1176.351057] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9955744b-2aea-43b6-b714-a0d13076ad57 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquired lock "refresh_cache-2d51997d-f88c-468c-9a56-0331a06ca56c" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1176.351344] env[63531]: DEBUG nova.network.neutron [None req-9955744b-2aea-43b6-b714-a0d13076ad57 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1176.351659] env[63531]: DEBUG nova.objects.instance [None req-9955744b-2aea-43b6-b714-a0d13076ad57 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lazy-loading 'info_cache' on Instance uuid 2d51997d-f88c-468c-9a56-0331a06ca56c {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1176.746149] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bedd6b36-c032-48e9-8bdb-dd1025e75dcf tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "ccc8f366-5849-45a6-9745-ff9b2be9c3bb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1176.746459] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bedd6b36-c032-48e9-8bdb-dd1025e75dcf tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "ccc8f366-5849-45a6-9745-ff9b2be9c3bb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1176.746697] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bedd6b36-c032-48e9-8bdb-dd1025e75dcf tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "ccc8f366-5849-45a6-9745-ff9b2be9c3bb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1176.746902] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bedd6b36-c032-48e9-8bdb-dd1025e75dcf tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "ccc8f366-5849-45a6-9745-ff9b2be9c3bb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1176.747203] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bedd6b36-c032-48e9-8bdb-dd1025e75dcf tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "ccc8f366-5849-45a6-9745-ff9b2be9c3bb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1176.749672] env[63531]: INFO nova.compute.manager [None req-bedd6b36-c032-48e9-8bdb-dd1025e75dcf tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Terminating instance [ 1176.751647] env[63531]: DEBUG nova.compute.manager [None req-bedd6b36-c032-48e9-8bdb-dd1025e75dcf tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1176.751861] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-bedd6b36-c032-48e9-8bdb-dd1025e75dcf tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1176.752753] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-012ca144-a590-4144-b587-c61143728011 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.760886] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-bedd6b36-c032-48e9-8bdb-dd1025e75dcf tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1176.761182] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-58d07bd7-ee77-45ae-9b50-f1dccbbeb4e1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.768229] env[63531]: DEBUG oslo_vmware.api [None req-bedd6b36-c032-48e9-8bdb-dd1025e75dcf tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for the task: (returnval){ [ 1176.768229] env[63531]: value = "task-1118753" [ 1176.768229] env[63531]: _type = "Task" [ 1176.768229] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1176.776187] env[63531]: DEBUG oslo_vmware.api [None req-bedd6b36-c032-48e9-8bdb-dd1025e75dcf tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118753, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.855925] env[63531]: DEBUG nova.objects.base [None req-9955744b-2aea-43b6-b714-a0d13076ad57 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Object Instance<2d51997d-f88c-468c-9a56-0331a06ca56c> lazy-loaded attributes: flavor,info_cache {{(pid=63531) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1177.281042] env[63531]: DEBUG oslo_vmware.api [None req-bedd6b36-c032-48e9-8bdb-dd1025e75dcf tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118753, 'name': PowerOffVM_Task, 'duration_secs': 0.202636} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1177.281042] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-bedd6b36-c032-48e9-8bdb-dd1025e75dcf tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1177.281248] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-bedd6b36-c032-48e9-8bdb-dd1025e75dcf tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1177.281484] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4b95f4fd-c79b-4297-97e4-1b45df59b499 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.346961] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-bedd6b36-c032-48e9-8bdb-dd1025e75dcf tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1177.347314] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-bedd6b36-c032-48e9-8bdb-dd1025e75dcf tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Deleting contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1177.347547] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-bedd6b36-c032-48e9-8bdb-dd1025e75dcf tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Deleting the datastore file [datastore2] ccc8f366-5849-45a6-9745-ff9b2be9c3bb {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1177.347828] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-378dd639-d4e6-4b85-9000-18ad13512590 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.354862] env[63531]: DEBUG oslo_vmware.api [None req-bedd6b36-c032-48e9-8bdb-dd1025e75dcf tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for the task: (returnval){ [ 1177.354862] env[63531]: value = "task-1118755" [ 1177.354862] env[63531]: _type = "Task" [ 1177.354862] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1177.364631] env[63531]: DEBUG oslo_vmware.api [None req-bedd6b36-c032-48e9-8bdb-dd1025e75dcf tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118755, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.593852] env[63531]: DEBUG nova.network.neutron [None req-9955744b-2aea-43b6-b714-a0d13076ad57 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Updating instance_info_cache with network_info: [{"id": "8c2307f2-c72e-47bb-a1f7-1dafecb0d2a4", "address": "fa:16:3e:54:75:66", "network": {"id": "0667c176-cfdf-4e43-b152-e00bdc224448", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-243715985-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.217", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "308eabafea9c4188a58a0f1c22074d2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8f580e6-1d86-41ee-9ebe-c531cb9299c6", "external-id": "nsx-vlan-transportzone-150", "segmentation_id": 150, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c2307f2-c7", "ovs_interfaceid": "8c2307f2-c72e-47bb-a1f7-1dafecb0d2a4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1177.865728] env[63531]: DEBUG oslo_vmware.api [None req-bedd6b36-c032-48e9-8bdb-dd1025e75dcf tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Task: {'id': task-1118755, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.120552} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1177.865969] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-bedd6b36-c032-48e9-8bdb-dd1025e75dcf tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1177.866223] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-bedd6b36-c032-48e9-8bdb-dd1025e75dcf tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Deleted contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1177.866455] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-bedd6b36-c032-48e9-8bdb-dd1025e75dcf tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1177.866654] env[63531]: INFO nova.compute.manager [None req-bedd6b36-c032-48e9-8bdb-dd1025e75dcf tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1177.866958] env[63531]: DEBUG oslo.service.loopingcall [None req-bedd6b36-c032-48e9-8bdb-dd1025e75dcf tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1177.867203] env[63531]: DEBUG nova.compute.manager [-] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1177.867302] env[63531]: DEBUG nova.network.neutron [-] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1178.097187] env[63531]: DEBUG oslo_concurrency.lockutils [None req-9955744b-2aea-43b6-b714-a0d13076ad57 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Releasing lock "refresh_cache-2d51997d-f88c-468c-9a56-0331a06ca56c" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1178.116275] env[63531]: DEBUG nova.compute.manager [req-7311220c-ae5e-48af-960b-57240676848e req-28c99cbe-658d-48e6-a351-74577df12d97 service nova] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Received event network-vif-deleted-e9692fb2-f3a9-455c-8317-b533a6af48f7 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1178.116401] env[63531]: INFO nova.compute.manager [req-7311220c-ae5e-48af-960b-57240676848e req-28c99cbe-658d-48e6-a351-74577df12d97 service nova] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Neutron deleted interface e9692fb2-f3a9-455c-8317-b533a6af48f7; detaching it from the instance and deleting it from the info cache [ 1178.116521] env[63531]: DEBUG nova.network.neutron [req-7311220c-ae5e-48af-960b-57240676848e req-28c99cbe-658d-48e6-a351-74577df12d97 service nova] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1178.593296] env[63531]: DEBUG nova.network.neutron [-] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1178.600774] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-9955744b-2aea-43b6-b714-a0d13076ad57 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1178.601208] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-13117034-3999-4a3e-8123-c8ab95221eb7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.615062] env[63531]: DEBUG oslo_vmware.api [None req-9955744b-2aea-43b6-b714-a0d13076ad57 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 1178.615062] env[63531]: value = "task-1118756" [ 1178.615062] env[63531]: _type = "Task" [ 1178.615062] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.620037] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2b8bb710-106b-4f99-a4cb-54dfe9524c63 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.630190] env[63531]: DEBUG oslo_vmware.api [None req-9955744b-2aea-43b6-b714-a0d13076ad57 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118756, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.636490] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b01499e-5615-4227-913e-32b6520a1c97 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.675296] env[63531]: DEBUG nova.compute.manager [req-7311220c-ae5e-48af-960b-57240676848e req-28c99cbe-658d-48e6-a351-74577df12d97 service nova] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Detach interface failed, port_id=e9692fb2-f3a9-455c-8317-b533a6af48f7, reason: Instance ccc8f366-5849-45a6-9745-ff9b2be9c3bb could not be found. {{(pid=63531) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1179.096586] env[63531]: INFO nova.compute.manager [-] [instance: ccc8f366-5849-45a6-9745-ff9b2be9c3bb] Took 1.23 seconds to deallocate network for instance. [ 1179.125657] env[63531]: DEBUG oslo_vmware.api [None req-9955744b-2aea-43b6-b714-a0d13076ad57 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118756, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.603747] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bedd6b36-c032-48e9-8bdb-dd1025e75dcf tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1179.604123] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bedd6b36-c032-48e9-8bdb-dd1025e75dcf tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1179.604343] env[63531]: DEBUG nova.objects.instance [None req-bedd6b36-c032-48e9-8bdb-dd1025e75dcf tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lazy-loading 'resources' on Instance uuid ccc8f366-5849-45a6-9745-ff9b2be9c3bb {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1179.627907] env[63531]: DEBUG oslo_vmware.api [None req-9955744b-2aea-43b6-b714-a0d13076ad57 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118756, 'name': PowerOnVM_Task, 'duration_secs': 0.813728} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1179.628616] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-9955744b-2aea-43b6-b714-a0d13076ad57 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1179.628791] env[63531]: DEBUG nova.compute.manager [None req-9955744b-2aea-43b6-b714-a0d13076ad57 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1179.629654] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45708c03-4a35-452f-8139-b91b0ac725a9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.169249] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-966678dc-61a7-4377-b278-cc5168b636db {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.177740] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bd67fa8-1013-46e8-9165-f465aef65524 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.224419] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c216e90-08c5-4299-b323-a2d307e7e743 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.234917] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c70305b4-caf4-49ba-8c5b-ba3634367146 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.255377] env[63531]: DEBUG nova.compute.provider_tree [None req-bedd6b36-c032-48e9-8bdb-dd1025e75dcf tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1180.759324] env[63531]: DEBUG nova.scheduler.client.report [None req-bedd6b36-c032-48e9-8bdb-dd1025e75dcf tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1180.912131] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d66f427-ff38-4aa3-9487-0b703c44f7e0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.919070] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-2d388e12-74d2-4a6d-a012-3884711b35cf tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Suspending the VM {{(pid=63531) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 1180.919321] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-ccb52886-5f74-4a5c-8f86-2d4a1af1a80f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.926578] env[63531]: DEBUG oslo_vmware.api [None req-2d388e12-74d2-4a6d-a012-3884711b35cf tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 1180.926578] env[63531]: value = "task-1118757" [ 1180.926578] env[63531]: _type = "Task" [ 1180.926578] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1180.934444] env[63531]: DEBUG oslo_vmware.api [None req-2d388e12-74d2-4a6d-a012-3884711b35cf tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118757, 'name': SuspendVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.264750] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bedd6b36-c032-48e9-8bdb-dd1025e75dcf tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.660s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1181.284206] env[63531]: INFO nova.scheduler.client.report [None req-bedd6b36-c032-48e9-8bdb-dd1025e75dcf tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Deleted allocations for instance ccc8f366-5849-45a6-9745-ff9b2be9c3bb [ 1181.437131] env[63531]: DEBUG oslo_vmware.api [None req-2d388e12-74d2-4a6d-a012-3884711b35cf tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118757, 'name': SuspendVM_Task} progress is 70%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.792670] env[63531]: DEBUG oslo_concurrency.lockutils [None req-bedd6b36-c032-48e9-8bdb-dd1025e75dcf tempest-AttachInterfacesTestJSON-1790236189 tempest-AttachInterfacesTestJSON-1790236189-project-member] Lock "ccc8f366-5849-45a6-9745-ff9b2be9c3bb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.046s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1181.938762] env[63531]: DEBUG oslo_vmware.api [None req-2d388e12-74d2-4a6d-a012-3884711b35cf tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118757, 'name': SuspendVM_Task, 'duration_secs': 0.705534} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1181.939089] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-2d388e12-74d2-4a6d-a012-3884711b35cf tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Suspended the VM {{(pid=63531) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 1181.939254] env[63531]: DEBUG nova.compute.manager [None req-2d388e12-74d2-4a6d-a012-3884711b35cf tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1181.940130] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9431c8b-40e3-441a-b1cb-1d7bf466a271 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.334020] env[63531]: DEBUG oslo_vmware.rw_handles [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b6948b-1438-2ad5-3304-dd9d39532d8d/disk-0.vmdk. {{(pid=63531) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1183.334020] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e84a66e2-b541-40d6-a5ed-2b6002b34688 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.340775] env[63531]: DEBUG oslo_vmware.rw_handles [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b6948b-1438-2ad5-3304-dd9d39532d8d/disk-0.vmdk is in state: ready. {{(pid=63531) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1183.341162] env[63531]: ERROR oslo_vmware.rw_handles [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b6948b-1438-2ad5-3304-dd9d39532d8d/disk-0.vmdk due to incomplete transfer. [ 1183.341526] env[63531]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-d702f0b0-84b3-4eb7-9353-a6f7347431e6 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.349178] env[63531]: DEBUG oslo_vmware.rw_handles [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b6948b-1438-2ad5-3304-dd9d39532d8d/disk-0.vmdk. {{(pid=63531) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1183.349522] env[63531]: DEBUG nova.virt.vmwareapi.images [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Uploaded image c6bead00-8690-442e-a30b-eb4e618e7b4b to the Glance image server {{(pid=63531) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1183.351977] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Destroying the VM {{(pid=63531) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1183.352351] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-a50ffd07-e786-4f88-97c5-0631ba754d3d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.359628] env[63531]: DEBUG oslo_vmware.api [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 1183.359628] env[63531]: value = "task-1118758" [ 1183.359628] env[63531]: _type = "Task" [ 1183.359628] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.367605] env[63531]: DEBUG oslo_vmware.api [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118758, 'name': Destroy_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.379805] env[63531]: INFO nova.compute.manager [None req-38db96a6-dc89-44fa-b9e6-5be3b734d978 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Resuming [ 1183.381279] env[63531]: DEBUG nova.objects.instance [None req-38db96a6-dc89-44fa-b9e6-5be3b734d978 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lazy-loading 'flavor' on Instance uuid 2d51997d-f88c-468c-9a56-0331a06ca56c {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1183.869990] env[63531]: DEBUG oslo_vmware.api [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118758, 'name': Destroy_Task, 'duration_secs': 0.331396} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1183.870343] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Destroyed the VM [ 1183.870640] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Deleting Snapshot of the VM instance {{(pid=63531) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1183.870925] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-6fb7cfc9-3ead-4b18-b195-0100f43e2f10 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.878444] env[63531]: DEBUG oslo_vmware.api [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 1183.878444] env[63531]: value = "task-1118759" [ 1183.878444] env[63531]: _type = "Task" [ 1183.878444] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.890048] env[63531]: DEBUG oslo_vmware.api [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118759, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.388883] env[63531]: DEBUG oslo_vmware.api [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118759, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.393013] env[63531]: DEBUG oslo_concurrency.lockutils [None req-38db96a6-dc89-44fa-b9e6-5be3b734d978 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "refresh_cache-2d51997d-f88c-468c-9a56-0331a06ca56c" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1184.393199] env[63531]: DEBUG oslo_concurrency.lockutils [None req-38db96a6-dc89-44fa-b9e6-5be3b734d978 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquired lock "refresh_cache-2d51997d-f88c-468c-9a56-0331a06ca56c" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1184.393387] env[63531]: DEBUG nova.network.neutron [None req-38db96a6-dc89-44fa-b9e6-5be3b734d978 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1184.893033] env[63531]: DEBUG oslo_vmware.api [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118759, 'name': RemoveSnapshot_Task, 'duration_secs': 0.731841} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.893306] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Deleted Snapshot of the VM instance {{(pid=63531) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1184.893592] env[63531]: DEBUG nova.compute.manager [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1184.894415] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc09b4ad-7e2d-437f-8ea7-53dd02c17382 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.142834] env[63531]: DEBUG nova.network.neutron [None req-38db96a6-dc89-44fa-b9e6-5be3b734d978 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Updating instance_info_cache with network_info: [{"id": "8c2307f2-c72e-47bb-a1f7-1dafecb0d2a4", "address": "fa:16:3e:54:75:66", "network": {"id": "0667c176-cfdf-4e43-b152-e00bdc224448", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-243715985-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.217", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "308eabafea9c4188a58a0f1c22074d2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8f580e6-1d86-41ee-9ebe-c531cb9299c6", "external-id": "nsx-vlan-transportzone-150", "segmentation_id": 150, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c2307f2-c7", "ovs_interfaceid": "8c2307f2-c72e-47bb-a1f7-1dafecb0d2a4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1185.407474] env[63531]: INFO nova.compute.manager [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Shelve offloading [ 1185.409542] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1185.409761] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b0b9b79a-ee63-4d56-958e-df9723acabaf {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.418290] env[63531]: DEBUG oslo_vmware.api [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 1185.418290] env[63531]: value = "task-1118760" [ 1185.418290] env[63531]: _type = "Task" [ 1185.418290] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.430008] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] VM already powered off {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1185.430278] env[63531]: DEBUG nova.compute.manager [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1185.431318] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-006ac8e4-9348-4c11-a714-81357a73143d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.438267] env[63531]: DEBUG oslo_concurrency.lockutils [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquiring lock "refresh_cache-3794f44c-8f6d-4bab-b9e0-9f6c819e1a97" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1185.438521] env[63531]: DEBUG oslo_concurrency.lockutils [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquired lock "refresh_cache-3794f44c-8f6d-4bab-b9e0-9f6c819e1a97" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1185.438757] env[63531]: DEBUG nova.network.neutron [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1185.649888] env[63531]: DEBUG oslo_concurrency.lockutils [None req-38db96a6-dc89-44fa-b9e6-5be3b734d978 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Releasing lock "refresh_cache-2d51997d-f88c-468c-9a56-0331a06ca56c" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1185.650932] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78841233-dde5-46d7-8f6e-7aad162b7592 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.658455] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-38db96a6-dc89-44fa-b9e6-5be3b734d978 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Resuming the VM {{(pid=63531) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1184}} [ 1185.658818] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-046cdd85-b1d4-4339-a199-2f2fa980d300 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.665242] env[63531]: DEBUG oslo_vmware.api [None req-38db96a6-dc89-44fa-b9e6-5be3b734d978 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 1185.665242] env[63531]: value = "task-1118761" [ 1185.665242] env[63531]: _type = "Task" [ 1185.665242] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.673902] env[63531]: DEBUG oslo_vmware.api [None req-38db96a6-dc89-44fa-b9e6-5be3b734d978 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118761, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.182543] env[63531]: DEBUG oslo_vmware.api [None req-38db96a6-dc89-44fa-b9e6-5be3b734d978 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118761, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.210126] env[63531]: DEBUG nova.network.neutron [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Updating instance_info_cache with network_info: [{"id": "9a3bf9ef-a105-4821-81f8-4891f7c3ae0a", "address": "fa:16:3e:dd:5e:bd", "network": {"id": "c4f4c994-36ab-42ab-94f8-78f29cbafa50", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-994507698-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "57e9bf4ee0f1463fa67b0f8deff30fdc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "767a3a48-41d4-4a0c-961d-0024837f63bd", "external-id": "nsx-vlan-transportzone-141", "segmentation_id": 141, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a3bf9ef-a1", "ovs_interfaceid": "9a3bf9ef-a105-4821-81f8-4891f7c3ae0a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1186.676426] env[63531]: DEBUG oslo_vmware.api [None req-38db96a6-dc89-44fa-b9e6-5be3b734d978 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118761, 'name': PowerOnVM_Task, 'duration_secs': 0.534222} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.676699] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-38db96a6-dc89-44fa-b9e6-5be3b734d978 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Resumed the VM {{(pid=63531) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1189}} [ 1186.676890] env[63531]: DEBUG nova.compute.manager [None req-38db96a6-dc89-44fa-b9e6-5be3b734d978 tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1186.677727] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87b37948-01d3-418d-a149-bf1edd94cfab {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.718263] env[63531]: DEBUG oslo_concurrency.lockutils [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Releasing lock "refresh_cache-3794f44c-8f6d-4bab-b9e0-9f6c819e1a97" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1186.979944] env[63531]: DEBUG nova.compute.manager [req-73ff4ae0-71a5-48ad-9436-56d8222af76b req-cbb7d1c9-6b2c-4e1a-b9c9-98dd11e47243 service nova] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Received event network-vif-unplugged-9a3bf9ef-a105-4821-81f8-4891f7c3ae0a {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1186.980190] env[63531]: DEBUG oslo_concurrency.lockutils [req-73ff4ae0-71a5-48ad-9436-56d8222af76b req-cbb7d1c9-6b2c-4e1a-b9c9-98dd11e47243 service nova] Acquiring lock "3794f44c-8f6d-4bab-b9e0-9f6c819e1a97-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1186.980405] env[63531]: DEBUG oslo_concurrency.lockutils [req-73ff4ae0-71a5-48ad-9436-56d8222af76b req-cbb7d1c9-6b2c-4e1a-b9c9-98dd11e47243 service nova] Lock "3794f44c-8f6d-4bab-b9e0-9f6c819e1a97-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1186.983432] env[63531]: DEBUG oslo_concurrency.lockutils [req-73ff4ae0-71a5-48ad-9436-56d8222af76b req-cbb7d1c9-6b2c-4e1a-b9c9-98dd11e47243 service nova] Lock "3794f44c-8f6d-4bab-b9e0-9f6c819e1a97-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1186.983432] env[63531]: DEBUG nova.compute.manager [req-73ff4ae0-71a5-48ad-9436-56d8222af76b req-cbb7d1c9-6b2c-4e1a-b9c9-98dd11e47243 service nova] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] No waiting events found dispatching network-vif-unplugged-9a3bf9ef-a105-4821-81f8-4891f7c3ae0a {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1186.983432] env[63531]: WARNING nova.compute.manager [req-73ff4ae0-71a5-48ad-9436-56d8222af76b req-cbb7d1c9-6b2c-4e1a-b9c9-98dd11e47243 service nova] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Received unexpected event network-vif-unplugged-9a3bf9ef-a105-4821-81f8-4891f7c3ae0a for instance with vm_state shelved and task_state shelving_offloading. [ 1187.079706] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1187.081496] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaab4428-36f2-488a-81ca-2e492ea7adfa {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.089101] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1187.089846] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1092f982-d300-4366-9602-1efb3329f4f7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.185633] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1187.186305] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Deleting contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1187.186305] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Deleting the datastore file [datastore2] 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1187.186552] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-298ba3ee-9775-4060-b673-95da8f0c5871 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.196154] env[63531]: DEBUG oslo_vmware.api [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 1187.196154] env[63531]: value = "task-1118763" [ 1187.196154] env[63531]: _type = "Task" [ 1187.196154] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.205369] env[63531]: DEBUG oslo_vmware.api [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118763, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.705924] env[63531]: DEBUG oslo_concurrency.lockutils [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Acquiring lock "7e32acbd-1996-4109-861a-d8f5d3177d5b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1187.706226] env[63531]: DEBUG oslo_concurrency.lockutils [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Lock "7e32acbd-1996-4109-861a-d8f5d3177d5b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1187.710822] env[63531]: DEBUG oslo_vmware.api [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118763, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141149} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1187.711467] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1187.711537] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Deleted contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1187.711767] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1187.732270] env[63531]: INFO nova.scheduler.client.report [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Deleted allocations for instance 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97 [ 1188.212463] env[63531]: DEBUG nova.compute.manager [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] [instance: 7e32acbd-1996-4109-861a-d8f5d3177d5b] Starting instance... {{(pid=63531) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1188.237520] env[63531]: DEBUG oslo_concurrency.lockutils [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1188.237794] env[63531]: DEBUG oslo_concurrency.lockutils [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1188.238033] env[63531]: DEBUG nova.objects.instance [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lazy-loading 'resources' on Instance uuid 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1188.310506] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e90a5a9e-3c5a-4b85-a786-ce7e15109c6a tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "2d51997d-f88c-468c-9a56-0331a06ca56c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1188.310785] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e90a5a9e-3c5a-4b85-a786-ce7e15109c6a tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "2d51997d-f88c-468c-9a56-0331a06ca56c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1188.310995] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e90a5a9e-3c5a-4b85-a786-ce7e15109c6a tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "2d51997d-f88c-468c-9a56-0331a06ca56c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1188.311202] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e90a5a9e-3c5a-4b85-a786-ce7e15109c6a tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "2d51997d-f88c-468c-9a56-0331a06ca56c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1188.311448] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e90a5a9e-3c5a-4b85-a786-ce7e15109c6a tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "2d51997d-f88c-468c-9a56-0331a06ca56c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1188.314074] env[63531]: INFO nova.compute.manager [None req-e90a5a9e-3c5a-4b85-a786-ce7e15109c6a tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Terminating instance [ 1188.315838] env[63531]: DEBUG nova.compute.manager [None req-e90a5a9e-3c5a-4b85-a786-ce7e15109c6a tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1188.316043] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e90a5a9e-3c5a-4b85-a786-ce7e15109c6a tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1188.317151] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c5a2126-f74a-4032-bfe4-6f8b219186d9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.325399] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-e90a5a9e-3c5a-4b85-a786-ce7e15109c6a tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1188.325749] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-263f34c0-b01f-486c-9e41-ded951826c58 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.332504] env[63531]: DEBUG oslo_vmware.api [None req-e90a5a9e-3c5a-4b85-a786-ce7e15109c6a tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 1188.332504] env[63531]: value = "task-1118764" [ 1188.332504] env[63531]: _type = "Task" [ 1188.332504] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.340080] env[63531]: DEBUG oslo_vmware.api [None req-e90a5a9e-3c5a-4b85-a786-ce7e15109c6a tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118764, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.734912] env[63531]: DEBUG oslo_concurrency.lockutils [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1188.741023] env[63531]: DEBUG nova.objects.instance [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lazy-loading 'numa_topology' on Instance uuid 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1188.844412] env[63531]: DEBUG oslo_vmware.api [None req-e90a5a9e-3c5a-4b85-a786-ce7e15109c6a tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118764, 'name': PowerOffVM_Task, 'duration_secs': 0.195693} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.844883] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-e90a5a9e-3c5a-4b85-a786-ce7e15109c6a tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1188.845153] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e90a5a9e-3c5a-4b85-a786-ce7e15109c6a tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1188.845433] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-826cbe82-4977-46c7-a046-16d3cc4dbc9d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.905167] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e90a5a9e-3c5a-4b85-a786-ce7e15109c6a tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1188.905167] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e90a5a9e-3c5a-4b85-a786-ce7e15109c6a tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Deleting contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1188.905167] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-e90a5a9e-3c5a-4b85-a786-ce7e15109c6a tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Deleting the datastore file [datastore2] 2d51997d-f88c-468c-9a56-0331a06ca56c {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1188.905292] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e6ea78a9-671f-4861-81e3-d8f3569ab19c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.911501] env[63531]: DEBUG oslo_vmware.api [None req-e90a5a9e-3c5a-4b85-a786-ce7e15109c6a tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for the task: (returnval){ [ 1188.911501] env[63531]: value = "task-1118766" [ 1188.911501] env[63531]: _type = "Task" [ 1188.911501] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.919464] env[63531]: DEBUG oslo_vmware.api [None req-e90a5a9e-3c5a-4b85-a786-ce7e15109c6a tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118766, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.074536] env[63531]: DEBUG nova.compute.manager [req-75864f6f-4c39-4f50-a123-44260af0ba25 req-1fe875b9-53f3-47d1-b80f-5f6a9e07f306 service nova] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Received event network-changed-9a3bf9ef-a105-4821-81f8-4891f7c3ae0a {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1189.074753] env[63531]: DEBUG nova.compute.manager [req-75864f6f-4c39-4f50-a123-44260af0ba25 req-1fe875b9-53f3-47d1-b80f-5f6a9e07f306 service nova] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Refreshing instance network info cache due to event network-changed-9a3bf9ef-a105-4821-81f8-4891f7c3ae0a. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1189.075013] env[63531]: DEBUG oslo_concurrency.lockutils [req-75864f6f-4c39-4f50-a123-44260af0ba25 req-1fe875b9-53f3-47d1-b80f-5f6a9e07f306 service nova] Acquiring lock "refresh_cache-3794f44c-8f6d-4bab-b9e0-9f6c819e1a97" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1189.075324] env[63531]: DEBUG oslo_concurrency.lockutils [req-75864f6f-4c39-4f50-a123-44260af0ba25 req-1fe875b9-53f3-47d1-b80f-5f6a9e07f306 service nova] Acquired lock "refresh_cache-3794f44c-8f6d-4bab-b9e0-9f6c819e1a97" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1189.075510] env[63531]: DEBUG nova.network.neutron [req-75864f6f-4c39-4f50-a123-44260af0ba25 req-1fe875b9-53f3-47d1-b80f-5f6a9e07f306 service nova] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Refreshing network info cache for port 9a3bf9ef-a105-4821-81f8-4891f7c3ae0a {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1189.243479] env[63531]: DEBUG nova.objects.base [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Object Instance<3794f44c-8f6d-4bab-b9e0-9f6c819e1a97> lazy-loaded attributes: resources,numa_topology {{(pid=63531) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1189.288255] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9510f6cf-e777-40e4-9e10-63696fb1afb8 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.296047] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d80bfe9b-5ce4-4d08-9bff-1cf4a560ef04 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.324904] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1291203-aad4-4604-8d7f-e476274d6894 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.331529] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91168ec3-6340-4ad8-ab42-9fd836756816 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.344735] env[63531]: DEBUG nova.compute.provider_tree [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1189.420855] env[63531]: DEBUG oslo_vmware.api [None req-e90a5a9e-3c5a-4b85-a786-ce7e15109c6a tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Task: {'id': task-1118766, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.138279} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1189.421137] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-e90a5a9e-3c5a-4b85-a786-ce7e15109c6a tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1189.421362] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e90a5a9e-3c5a-4b85-a786-ce7e15109c6a tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Deleted contents of the VM from datastore datastore2 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1189.421563] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-e90a5a9e-3c5a-4b85-a786-ce7e15109c6a tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1189.421740] env[63531]: INFO nova.compute.manager [None req-e90a5a9e-3c5a-4b85-a786-ce7e15109c6a tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1189.421980] env[63531]: DEBUG oslo.service.loopingcall [None req-e90a5a9e-3c5a-4b85-a786-ce7e15109c6a tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1189.422185] env[63531]: DEBUG nova.compute.manager [-] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1189.422284] env[63531]: DEBUG nova.network.neutron [-] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1190.298904] env[63531]: DEBUG nova.scheduler.client.report [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1190.301897] env[63531]: DEBUG oslo_concurrency.lockutils [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquiring lock "3794f44c-8f6d-4bab-b9e0-9f6c819e1a97" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1190.498978] env[63531]: DEBUG nova.network.neutron [req-75864f6f-4c39-4f50-a123-44260af0ba25 req-1fe875b9-53f3-47d1-b80f-5f6a9e07f306 service nova] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Updated VIF entry in instance network info cache for port 9a3bf9ef-a105-4821-81f8-4891f7c3ae0a. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1190.499453] env[63531]: DEBUG nova.network.neutron [req-75864f6f-4c39-4f50-a123-44260af0ba25 req-1fe875b9-53f3-47d1-b80f-5f6a9e07f306 service nova] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Updating instance_info_cache with network_info: [{"id": "9a3bf9ef-a105-4821-81f8-4891f7c3ae0a", "address": "fa:16:3e:dd:5e:bd", "network": {"id": "c4f4c994-36ab-42ab-94f8-78f29cbafa50", "bridge": null, "label": "tempest-ServerActionsTestOtherB-994507698-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "57e9bf4ee0f1463fa67b0f8deff30fdc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap9a3bf9ef-a1", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1190.803104] env[63531]: DEBUG nova.network.neutron [-] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1190.804885] env[63531]: DEBUG oslo_concurrency.lockutils [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.567s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1190.807773] env[63531]: DEBUG oslo_concurrency.lockutils [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.073s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1190.809181] env[63531]: INFO nova.compute.claims [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] [instance: 7e32acbd-1996-4109-861a-d8f5d3177d5b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1191.002816] env[63531]: DEBUG oslo_concurrency.lockutils [req-75864f6f-4c39-4f50-a123-44260af0ba25 req-1fe875b9-53f3-47d1-b80f-5f6a9e07f306 service nova] Releasing lock "refresh_cache-3794f44c-8f6d-4bab-b9e0-9f6c819e1a97" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1191.100136] env[63531]: DEBUG nova.compute.manager [req-178a48d6-2e50-4a11-b125-a70d7e7a34ae req-2d21c1d1-5482-49ac-b350-03c17dd804ea service nova] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Received event network-vif-deleted-8c2307f2-c72e-47bb-a1f7-1dafecb0d2a4 {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1191.310126] env[63531]: INFO nova.compute.manager [-] [instance: 2d51997d-f88c-468c-9a56-0331a06ca56c] Took 1.89 seconds to deallocate network for instance. [ 1191.314797] env[63531]: DEBUG oslo_concurrency.lockutils [None req-45a5835e-77b8-4747-94a2-8e2684dc45f0 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "3794f44c-8f6d-4bab-b9e0-9f6c819e1a97" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 21.471s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1191.317865] env[63531]: DEBUG oslo_concurrency.lockutils [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "3794f44c-8f6d-4bab-b9e0-9f6c819e1a97" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 1.016s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1191.318085] env[63531]: INFO nova.compute.manager [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Unshelving [ 1191.825961] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e90a5a9e-3c5a-4b85-a786-ce7e15109c6a tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1191.881049] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb176322-3e1d-411f-a70d-87f862ffdf8b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.888219] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62624d79-e75c-4279-a42d-dbb7f2c2ecd7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.918140] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-430afa69-24ea-46ce-ada4-eeb9495331a7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.925126] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cc48ef8-222f-4433-ba32-98a5b4e73462 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.937840] env[63531]: DEBUG nova.compute.provider_tree [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1192.342956] env[63531]: DEBUG oslo_concurrency.lockutils [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1192.441749] env[63531]: DEBUG nova.scheduler.client.report [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1192.947637] env[63531]: DEBUG oslo_concurrency.lockutils [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.140s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1192.948238] env[63531]: DEBUG nova.compute.manager [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] [instance: 7e32acbd-1996-4109-861a-d8f5d3177d5b] Start building networks asynchronously for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1192.951218] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e90a5a9e-3c5a-4b85-a786-ce7e15109c6a tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.125s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1192.951487] env[63531]: DEBUG nova.objects.instance [None req-e90a5a9e-3c5a-4b85-a786-ce7e15109c6a tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lazy-loading 'resources' on Instance uuid 2d51997d-f88c-468c-9a56-0331a06ca56c {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1193.454857] env[63531]: DEBUG nova.compute.utils [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Using /dev/sd instead of None {{(pid=63531) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1193.458998] env[63531]: DEBUG nova.compute.manager [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] [instance: 7e32acbd-1996-4109-861a-d8f5d3177d5b] Allocating IP information in the background. {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1193.459189] env[63531]: DEBUG nova.network.neutron [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] [instance: 7e32acbd-1996-4109-861a-d8f5d3177d5b] allocate_for_instance() {{(pid=63531) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1193.502597] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-320e3c5b-1735-4dff-a3b2-8da26246ea67 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.510708] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae691d00-d42f-4c65-8f2f-c8e529c21d61 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.541180] env[63531]: DEBUG nova.policy [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '57ce2446ade545fb97b700f2a2d23fe8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6811b83d984b4c25b3b7da98a3421189', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63531) authorize /opt/stack/nova/nova/policy.py:201}} [ 1193.543025] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35f60174-bbb5-43c1-a24f-42e621b1e177 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.549923] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53c27937-85a5-4bde-9dfb-7eb723a83191 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.562544] env[63531]: DEBUG nova.compute.provider_tree [None req-e90a5a9e-3c5a-4b85-a786-ce7e15109c6a tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1193.800019] env[63531]: DEBUG nova.network.neutron [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] [instance: 7e32acbd-1996-4109-861a-d8f5d3177d5b] Successfully created port: 06ef1d52-c054-4224-9227-402b6a036abc {{(pid=63531) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1193.959698] env[63531]: DEBUG nova.compute.manager [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] [instance: 7e32acbd-1996-4109-861a-d8f5d3177d5b] Start building block device mappings for instance. {{(pid=63531) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1194.065619] env[63531]: DEBUG nova.scheduler.client.report [None req-e90a5a9e-3c5a-4b85-a786-ce7e15109c6a tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1194.570560] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e90a5a9e-3c5a-4b85-a786-ce7e15109c6a tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.619s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1194.572895] env[63531]: DEBUG oslo_concurrency.lockutils [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.230s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1194.573147] env[63531]: DEBUG nova.objects.instance [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lazy-loading 'pci_requests' on Instance uuid 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1194.593425] env[63531]: INFO nova.scheduler.client.report [None req-e90a5a9e-3c5a-4b85-a786-ce7e15109c6a tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Deleted allocations for instance 2d51997d-f88c-468c-9a56-0331a06ca56c [ 1194.969320] env[63531]: DEBUG nova.compute.manager [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] [instance: 7e32acbd-1996-4109-861a-d8f5d3177d5b] Start spawning the instance on the hypervisor. {{(pid=63531) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1194.993702] env[63531]: DEBUG nova.virt.hardware [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T08:14:27Z,direct_url=,disk_format='vmdk',id=9d550399-588c-4bd5-8436-eba9b523bd13,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='54fbef8a6ee0488f820706d2036e9c99',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T08:14:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1194.993955] env[63531]: DEBUG nova.virt.hardware [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1194.994140] env[63531]: DEBUG nova.virt.hardware [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1194.994335] env[63531]: DEBUG nova.virt.hardware [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1194.994510] env[63531]: DEBUG nova.virt.hardware [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1194.994670] env[63531]: DEBUG nova.virt.hardware [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1194.994886] env[63531]: DEBUG nova.virt.hardware [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1194.995065] env[63531]: DEBUG nova.virt.hardware [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1194.995240] env[63531]: DEBUG nova.virt.hardware [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1194.995411] env[63531]: DEBUG nova.virt.hardware [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1194.995592] env[63531]: DEBUG nova.virt.hardware [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1194.996460] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14db8788-88d5-4139-9f46-a3de74464ccd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.004409] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5d2ae7c-da9c-4c8b-af6c-3962653d1845 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.077150] env[63531]: DEBUG nova.objects.instance [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lazy-loading 'numa_topology' on Instance uuid 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1195.100316] env[63531]: DEBUG oslo_concurrency.lockutils [None req-e90a5a9e-3c5a-4b85-a786-ce7e15109c6a tempest-ServerActionsTestJSON-1369178052 tempest-ServerActionsTestJSON-1369178052-project-member] Lock "2d51997d-f88c-468c-9a56-0331a06ca56c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.789s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1195.210537] env[63531]: DEBUG nova.compute.manager [req-eea60d50-afd8-47ef-afb8-4c375596b13b req-9a6ad45a-f5ec-4710-be8b-f10492ac3b01 service nova] [instance: 7e32acbd-1996-4109-861a-d8f5d3177d5b] Received event network-vif-plugged-06ef1d52-c054-4224-9227-402b6a036abc {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1195.210761] env[63531]: DEBUG oslo_concurrency.lockutils [req-eea60d50-afd8-47ef-afb8-4c375596b13b req-9a6ad45a-f5ec-4710-be8b-f10492ac3b01 service nova] Acquiring lock "7e32acbd-1996-4109-861a-d8f5d3177d5b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1195.210967] env[63531]: DEBUG oslo_concurrency.lockutils [req-eea60d50-afd8-47ef-afb8-4c375596b13b req-9a6ad45a-f5ec-4710-be8b-f10492ac3b01 service nova] Lock "7e32acbd-1996-4109-861a-d8f5d3177d5b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1195.211187] env[63531]: DEBUG oslo_concurrency.lockutils [req-eea60d50-afd8-47ef-afb8-4c375596b13b req-9a6ad45a-f5ec-4710-be8b-f10492ac3b01 service nova] Lock "7e32acbd-1996-4109-861a-d8f5d3177d5b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1195.211323] env[63531]: DEBUG nova.compute.manager [req-eea60d50-afd8-47ef-afb8-4c375596b13b req-9a6ad45a-f5ec-4710-be8b-f10492ac3b01 service nova] [instance: 7e32acbd-1996-4109-861a-d8f5d3177d5b] No waiting events found dispatching network-vif-plugged-06ef1d52-c054-4224-9227-402b6a036abc {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1195.211602] env[63531]: WARNING nova.compute.manager [req-eea60d50-afd8-47ef-afb8-4c375596b13b req-9a6ad45a-f5ec-4710-be8b-f10492ac3b01 service nova] [instance: 7e32acbd-1996-4109-861a-d8f5d3177d5b] Received unexpected event network-vif-plugged-06ef1d52-c054-4224-9227-402b6a036abc for instance with vm_state building and task_state spawning. [ 1195.263534] env[63531]: DEBUG nova.network.neutron [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] [instance: 7e32acbd-1996-4109-861a-d8f5d3177d5b] Successfully updated port: 06ef1d52-c054-4224-9227-402b6a036abc {{(pid=63531) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1195.579866] env[63531]: INFO nova.compute.claims [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1195.766140] env[63531]: DEBUG oslo_concurrency.lockutils [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Acquiring lock "refresh_cache-7e32acbd-1996-4109-861a-d8f5d3177d5b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1195.766383] env[63531]: DEBUG oslo_concurrency.lockutils [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Acquired lock "refresh_cache-7e32acbd-1996-4109-861a-d8f5d3177d5b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1195.766617] env[63531]: DEBUG nova.network.neutron [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] [instance: 7e32acbd-1996-4109-861a-d8f5d3177d5b] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1196.314260] env[63531]: DEBUG nova.network.neutron [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] [instance: 7e32acbd-1996-4109-861a-d8f5d3177d5b] Instance cache missing network info. {{(pid=63531) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1196.529369] env[63531]: DEBUG nova.network.neutron [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] [instance: 7e32acbd-1996-4109-861a-d8f5d3177d5b] Updating instance_info_cache with network_info: [{"id": "06ef1d52-c054-4224-9227-402b6a036abc", "address": "fa:16:3e:2d:c1:a9", "network": {"id": "3dab3f08-5511-4014-818e-36b3196ebba5", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-937948863-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6811b83d984b4c25b3b7da98a3421189", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4349e30-c086-4c24-9e0e-83996d808a1b", "external-id": "nsx-vlan-transportzone-266", "segmentation_id": 266, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap06ef1d52-c0", "ovs_interfaceid": "06ef1d52-c054-4224-9227-402b6a036abc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1196.629955] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aabb7ada-d763-4d81-b640-181ca995072b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.638165] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-982054c1-e62e-4a0e-b976-bbc4f907f716 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.667635] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f00198ad-9a94-4a47-b8ec-61d4031cc78c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.675364] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c4b55b4-84ff-45ff-856c-0c80785adf3e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.688388] env[63531]: DEBUG nova.compute.provider_tree [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1197.032093] env[63531]: DEBUG oslo_concurrency.lockutils [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Releasing lock "refresh_cache-7e32acbd-1996-4109-861a-d8f5d3177d5b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1197.032547] env[63531]: DEBUG nova.compute.manager [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] [instance: 7e32acbd-1996-4109-861a-d8f5d3177d5b] Instance network_info: |[{"id": "06ef1d52-c054-4224-9227-402b6a036abc", "address": "fa:16:3e:2d:c1:a9", "network": {"id": "3dab3f08-5511-4014-818e-36b3196ebba5", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-937948863-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6811b83d984b4c25b3b7da98a3421189", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4349e30-c086-4c24-9e0e-83996d808a1b", "external-id": "nsx-vlan-transportzone-266", "segmentation_id": 266, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap06ef1d52-c0", "ovs_interfaceid": "06ef1d52-c054-4224-9227-402b6a036abc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63531) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1197.033053] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] [instance: 7e32acbd-1996-4109-861a-d8f5d3177d5b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2d:c1:a9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c4349e30-c086-4c24-9e0e-83996d808a1b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '06ef1d52-c054-4224-9227-402b6a036abc', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1197.041167] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Creating folder: Project (6811b83d984b4c25b3b7da98a3421189). Parent ref: group-v244585. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1197.041494] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bcce2b27-8cb2-41f6-81da-6618224863d7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.055706] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Created folder: Project (6811b83d984b4c25b3b7da98a3421189) in parent group-v244585. [ 1197.055910] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Creating folder: Instances. Parent ref: group-v244821. {{(pid=63531) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1197.056158] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c202f4ba-9b9a-4ce5-b140-2017e09cf89d {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.065579] env[63531]: INFO nova.virt.vmwareapi.vm_util [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Created folder: Instances in parent group-v244821. [ 1197.065813] env[63531]: DEBUG oslo.service.loopingcall [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1197.066035] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7e32acbd-1996-4109-861a-d8f5d3177d5b] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1197.066465] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-33615ce5-963a-42b6-ad44-c8a2952bec87 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.084301] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1197.084301] env[63531]: value = "task-1118769" [ 1197.084301] env[63531]: _type = "Task" [ 1197.084301] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1197.091444] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118769, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.191890] env[63531]: DEBUG nova.scheduler.client.report [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1197.277788] env[63531]: DEBUG nova.compute.manager [req-00dc0331-3432-4628-a54d-9ba1382ad996 req-9ebf04bc-78bc-458f-a4a7-63dfec43089f service nova] [instance: 7e32acbd-1996-4109-861a-d8f5d3177d5b] Received event network-changed-06ef1d52-c054-4224-9227-402b6a036abc {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1197.277995] env[63531]: DEBUG nova.compute.manager [req-00dc0331-3432-4628-a54d-9ba1382ad996 req-9ebf04bc-78bc-458f-a4a7-63dfec43089f service nova] [instance: 7e32acbd-1996-4109-861a-d8f5d3177d5b] Refreshing instance network info cache due to event network-changed-06ef1d52-c054-4224-9227-402b6a036abc. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1197.278228] env[63531]: DEBUG oslo_concurrency.lockutils [req-00dc0331-3432-4628-a54d-9ba1382ad996 req-9ebf04bc-78bc-458f-a4a7-63dfec43089f service nova] Acquiring lock "refresh_cache-7e32acbd-1996-4109-861a-d8f5d3177d5b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1197.278376] env[63531]: DEBUG oslo_concurrency.lockutils [req-00dc0331-3432-4628-a54d-9ba1382ad996 req-9ebf04bc-78bc-458f-a4a7-63dfec43089f service nova] Acquired lock "refresh_cache-7e32acbd-1996-4109-861a-d8f5d3177d5b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1197.278782] env[63531]: DEBUG nova.network.neutron [req-00dc0331-3432-4628-a54d-9ba1382ad996 req-9ebf04bc-78bc-458f-a4a7-63dfec43089f service nova] [instance: 7e32acbd-1996-4109-861a-d8f5d3177d5b] Refreshing network info cache for port 06ef1d52-c054-4224-9227-402b6a036abc {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1197.594189] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118769, 'name': CreateVM_Task, 'duration_secs': 0.309524} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1197.594360] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7e32acbd-1996-4109-861a-d8f5d3177d5b] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1197.595035] env[63531]: DEBUG oslo_concurrency.lockutils [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1197.595207] env[63531]: DEBUG oslo_concurrency.lockutils [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1197.595525] env[63531]: DEBUG oslo_concurrency.lockutils [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1197.595769] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-77890734-11dc-4771-85a8-6cf176882cbf {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.599874] env[63531]: DEBUG oslo_vmware.api [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Waiting for the task: (returnval){ [ 1197.599874] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52275721-a806-5234-1130-5f18ed3ec125" [ 1197.599874] env[63531]: _type = "Task" [ 1197.599874] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1197.607321] env[63531]: DEBUG oslo_vmware.api [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52275721-a806-5234-1130-5f18ed3ec125, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.697469] env[63531]: DEBUG oslo_concurrency.lockutils [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.124s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1197.726924] env[63531]: INFO nova.network.neutron [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Updating port 9a3bf9ef-a105-4821-81f8-4891f7c3ae0a with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1197.972301] env[63531]: DEBUG nova.network.neutron [req-00dc0331-3432-4628-a54d-9ba1382ad996 req-9ebf04bc-78bc-458f-a4a7-63dfec43089f service nova] [instance: 7e32acbd-1996-4109-861a-d8f5d3177d5b] Updated VIF entry in instance network info cache for port 06ef1d52-c054-4224-9227-402b6a036abc. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1197.972692] env[63531]: DEBUG nova.network.neutron [req-00dc0331-3432-4628-a54d-9ba1382ad996 req-9ebf04bc-78bc-458f-a4a7-63dfec43089f service nova] [instance: 7e32acbd-1996-4109-861a-d8f5d3177d5b] Updating instance_info_cache with network_info: [{"id": "06ef1d52-c054-4224-9227-402b6a036abc", "address": "fa:16:3e:2d:c1:a9", "network": {"id": "3dab3f08-5511-4014-818e-36b3196ebba5", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-937948863-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6811b83d984b4c25b3b7da98a3421189", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4349e30-c086-4c24-9e0e-83996d808a1b", "external-id": "nsx-vlan-transportzone-266", "segmentation_id": 266, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap06ef1d52-c0", "ovs_interfaceid": "06ef1d52-c054-4224-9227-402b6a036abc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1198.110590] env[63531]: DEBUG oslo_vmware.api [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52275721-a806-5234-1130-5f18ed3ec125, 'name': SearchDatastore_Task, 'duration_secs': 0.01026} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1198.110903] env[63531]: DEBUG oslo_concurrency.lockutils [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1198.111166] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] [instance: 7e32acbd-1996-4109-861a-d8f5d3177d5b] Processing image 9d550399-588c-4bd5-8436-eba9b523bd13 {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1198.111496] env[63531]: DEBUG oslo_concurrency.lockutils [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1198.111588] env[63531]: DEBUG oslo_concurrency.lockutils [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1198.111780] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1198.112050] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6d64a304-921d-4dd7-84a4-6ce3cde7d335 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.120656] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1198.120844] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1198.121623] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1a58c644-e214-4252-8835-aed387e21466 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.126608] env[63531]: DEBUG oslo_vmware.api [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Waiting for the task: (returnval){ [ 1198.126608] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52b8ea9c-c62f-95a4-41c8-455853301c00" [ 1198.126608] env[63531]: _type = "Task" [ 1198.126608] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1198.134212] env[63531]: DEBUG oslo_vmware.api [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52b8ea9c-c62f-95a4-41c8-455853301c00, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1198.475128] env[63531]: DEBUG oslo_concurrency.lockutils [req-00dc0331-3432-4628-a54d-9ba1382ad996 req-9ebf04bc-78bc-458f-a4a7-63dfec43089f service nova] Releasing lock "refresh_cache-7e32acbd-1996-4109-861a-d8f5d3177d5b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1198.636770] env[63531]: DEBUG oslo_vmware.api [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]52b8ea9c-c62f-95a4-41c8-455853301c00, 'name': SearchDatastore_Task, 'duration_secs': 0.008457} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1198.637544] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c96ed0e8-b542-4454-99cd-c4eec7dae4a7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.642470] env[63531]: DEBUG oslo_vmware.api [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Waiting for the task: (returnval){ [ 1198.642470] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]527ec7e8-fef3-9066-254f-a2ebb6be3943" [ 1198.642470] env[63531]: _type = "Task" [ 1198.642470] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1198.649504] env[63531]: DEBUG oslo_vmware.api [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]527ec7e8-fef3-9066-254f-a2ebb6be3943, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1199.154779] env[63531]: DEBUG oslo_vmware.api [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]527ec7e8-fef3-9066-254f-a2ebb6be3943, 'name': SearchDatastore_Task, 'duration_secs': 0.009379} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1199.155058] env[63531]: DEBUG oslo_concurrency.lockutils [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1199.155268] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 7e32acbd-1996-4109-861a-d8f5d3177d5b/7e32acbd-1996-4109-861a-d8f5d3177d5b.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1199.155725] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-27d0b749-8676-425b-9519-66656281f527 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.161849] env[63531]: DEBUG oslo_vmware.api [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Waiting for the task: (returnval){ [ 1199.161849] env[63531]: value = "task-1118770" [ 1199.161849] env[63531]: _type = "Task" [ 1199.161849] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1199.169237] env[63531]: DEBUG oslo_vmware.api [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Task: {'id': task-1118770, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1199.185970] env[63531]: DEBUG oslo_concurrency.lockutils [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquiring lock "refresh_cache-3794f44c-8f6d-4bab-b9e0-9f6c819e1a97" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1199.186163] env[63531]: DEBUG oslo_concurrency.lockutils [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquired lock "refresh_cache-3794f44c-8f6d-4bab-b9e0-9f6c819e1a97" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1199.186372] env[63531]: DEBUG nova.network.neutron [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Building network info cache for instance {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1199.306522] env[63531]: DEBUG nova.compute.manager [req-f5689a88-bac4-48ef-bf1b-ec6bf6dff669 req-a5fa2671-f2ea-4da3-9c4f-a59ad039d3e9 service nova] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Received event network-vif-plugged-9a3bf9ef-a105-4821-81f8-4891f7c3ae0a {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1199.306921] env[63531]: DEBUG oslo_concurrency.lockutils [req-f5689a88-bac4-48ef-bf1b-ec6bf6dff669 req-a5fa2671-f2ea-4da3-9c4f-a59ad039d3e9 service nova] Acquiring lock "3794f44c-8f6d-4bab-b9e0-9f6c819e1a97-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1199.307072] env[63531]: DEBUG oslo_concurrency.lockutils [req-f5689a88-bac4-48ef-bf1b-ec6bf6dff669 req-a5fa2671-f2ea-4da3-9c4f-a59ad039d3e9 service nova] Lock "3794f44c-8f6d-4bab-b9e0-9f6c819e1a97-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1199.307309] env[63531]: DEBUG oslo_concurrency.lockutils [req-f5689a88-bac4-48ef-bf1b-ec6bf6dff669 req-a5fa2671-f2ea-4da3-9c4f-a59ad039d3e9 service nova] Lock "3794f44c-8f6d-4bab-b9e0-9f6c819e1a97-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1199.307497] env[63531]: DEBUG nova.compute.manager [req-f5689a88-bac4-48ef-bf1b-ec6bf6dff669 req-a5fa2671-f2ea-4da3-9c4f-a59ad039d3e9 service nova] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] No waiting events found dispatching network-vif-plugged-9a3bf9ef-a105-4821-81f8-4891f7c3ae0a {{(pid=63531) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1199.307701] env[63531]: WARNING nova.compute.manager [req-f5689a88-bac4-48ef-bf1b-ec6bf6dff669 req-a5fa2671-f2ea-4da3-9c4f-a59ad039d3e9 service nova] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Received unexpected event network-vif-plugged-9a3bf9ef-a105-4821-81f8-4891f7c3ae0a for instance with vm_state shelved_offloaded and task_state spawning. [ 1199.307893] env[63531]: DEBUG nova.compute.manager [req-f5689a88-bac4-48ef-bf1b-ec6bf6dff669 req-a5fa2671-f2ea-4da3-9c4f-a59ad039d3e9 service nova] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Received event network-changed-9a3bf9ef-a105-4821-81f8-4891f7c3ae0a {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1199.308078] env[63531]: DEBUG nova.compute.manager [req-f5689a88-bac4-48ef-bf1b-ec6bf6dff669 req-a5fa2671-f2ea-4da3-9c4f-a59ad039d3e9 service nova] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Refreshing instance network info cache due to event network-changed-9a3bf9ef-a105-4821-81f8-4891f7c3ae0a. {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1199.308254] env[63531]: DEBUG oslo_concurrency.lockutils [req-f5689a88-bac4-48ef-bf1b-ec6bf6dff669 req-a5fa2671-f2ea-4da3-9c4f-a59ad039d3e9 service nova] Acquiring lock "refresh_cache-3794f44c-8f6d-4bab-b9e0-9f6c819e1a97" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1199.671386] env[63531]: DEBUG oslo_vmware.api [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Task: {'id': task-1118770, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.453793} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1199.671802] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9d550399-588c-4bd5-8436-eba9b523bd13/9d550399-588c-4bd5-8436-eba9b523bd13.vmdk to [datastore1] 7e32acbd-1996-4109-861a-d8f5d3177d5b/7e32acbd-1996-4109-861a-d8f5d3177d5b.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1199.672052] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] [instance: 7e32acbd-1996-4109-861a-d8f5d3177d5b] Extending root virtual disk to 1048576 {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1199.672304] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3ea7f6c9-83c9-47d3-bbe4-9f1458310b23 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.679093] env[63531]: DEBUG oslo_vmware.api [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Waiting for the task: (returnval){ [ 1199.679093] env[63531]: value = "task-1118771" [ 1199.679093] env[63531]: _type = "Task" [ 1199.679093] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1199.686372] env[63531]: DEBUG oslo_vmware.api [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Task: {'id': task-1118771, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1199.929354] env[63531]: DEBUG nova.network.neutron [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Updating instance_info_cache with network_info: [{"id": "9a3bf9ef-a105-4821-81f8-4891f7c3ae0a", "address": "fa:16:3e:dd:5e:bd", "network": {"id": "c4f4c994-36ab-42ab-94f8-78f29cbafa50", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-994507698-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "57e9bf4ee0f1463fa67b0f8deff30fdc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "767a3a48-41d4-4a0c-961d-0024837f63bd", "external-id": "nsx-vlan-transportzone-141", "segmentation_id": 141, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a3bf9ef-a1", "ovs_interfaceid": "9a3bf9ef-a105-4821-81f8-4891f7c3ae0a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1200.188743] env[63531]: DEBUG oslo_vmware.api [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Task: {'id': task-1118771, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071562} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1200.189116] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] [instance: 7e32acbd-1996-4109-861a-d8f5d3177d5b] Extended root virtual disk {{(pid=63531) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1200.189781] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-189ef39c-839f-4425-84fa-e03f92f60d09 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.210819] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] [instance: 7e32acbd-1996-4109-861a-d8f5d3177d5b] Reconfiguring VM instance instance-00000075 to attach disk [datastore1] 7e32acbd-1996-4109-861a-d8f5d3177d5b/7e32acbd-1996-4109-861a-d8f5d3177d5b.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1200.211068] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4f2ab36d-48ba-4cbb-bd75-8422c62665f7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.230240] env[63531]: DEBUG oslo_vmware.api [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Waiting for the task: (returnval){ [ 1200.230240] env[63531]: value = "task-1118772" [ 1200.230240] env[63531]: _type = "Task" [ 1200.230240] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1200.238673] env[63531]: DEBUG oslo_vmware.api [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Task: {'id': task-1118772, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.432463] env[63531]: DEBUG oslo_concurrency.lockutils [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Releasing lock "refresh_cache-3794f44c-8f6d-4bab-b9e0-9f6c819e1a97" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1200.435041] env[63531]: DEBUG oslo_concurrency.lockutils [req-f5689a88-bac4-48ef-bf1b-ec6bf6dff669 req-a5fa2671-f2ea-4da3-9c4f-a59ad039d3e9 service nova] Acquired lock "refresh_cache-3794f44c-8f6d-4bab-b9e0-9f6c819e1a97" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1200.435283] env[63531]: DEBUG nova.network.neutron [req-f5689a88-bac4-48ef-bf1b-ec6bf6dff669 req-a5fa2671-f2ea-4da3-9c4f-a59ad039d3e9 service nova] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Refreshing network info cache for port 9a3bf9ef-a105-4821-81f8-4891f7c3ae0a {{(pid=63531) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1200.463140] env[63531]: DEBUG nova.virt.hardware [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T08:14:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='59f1551da2dd0bcb569d815dc6b3a65d',container_format='bare',created_at=2024-10-03T08:26:46Z,direct_url=,disk_format='vmdk',id=c6bead00-8690-442e-a30b-eb4e618e7b4b,min_disk=1,min_ram=0,name='tempest-ServerActionsTestOtherB-server-456121659-shelved',owner='57e9bf4ee0f1463fa67b0f8deff30fdc',properties=ImageMetaProps,protected=,size=31667712,status='active',tags=,updated_at=2024-10-03T08:27:00Z,virtual_size=,visibility=), allow threads: False {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1200.463531] env[63531]: DEBUG nova.virt.hardware [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Flavor limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1200.463757] env[63531]: DEBUG nova.virt.hardware [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Image limits 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1200.464079] env[63531]: DEBUG nova.virt.hardware [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Flavor pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1200.464337] env[63531]: DEBUG nova.virt.hardware [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Image pref 0:0:0 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1200.464523] env[63531]: DEBUG nova.virt.hardware [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63531) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1200.464825] env[63531]: DEBUG nova.virt.hardware [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1200.465047] env[63531]: DEBUG nova.virt.hardware [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1200.465236] env[63531]: DEBUG nova.virt.hardware [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Got 1 possible topologies {{(pid=63531) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1200.465411] env[63531]: DEBUG nova.virt.hardware [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1200.465590] env[63531]: DEBUG nova.virt.hardware [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63531) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1200.466640] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f361f1b1-eab4-42d8-b268-42c54f4644d7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.473935] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b912162a-6758-4b35-93c1-b9a97c316e2c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.487113] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:dd:5e:bd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '767a3a48-41d4-4a0c-961d-0024837f63bd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9a3bf9ef-a105-4821-81f8-4891f7c3ae0a', 'vif_model': 'vmxnet3'}] {{(pid=63531) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1200.494227] env[63531]: DEBUG oslo.service.loopingcall [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1200.494441] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Creating VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1200.494639] env[63531]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3413bceb-399c-4402-af80-0473b4634308 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.513321] env[63531]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1200.513321] env[63531]: value = "task-1118773" [ 1200.513321] env[63531]: _type = "Task" [ 1200.513321] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1200.520212] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118773, 'name': CreateVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.739798] env[63531]: DEBUG oslo_vmware.api [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Task: {'id': task-1118772, 'name': ReconfigVM_Task, 'duration_secs': 0.258067} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1200.740126] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] [instance: 7e32acbd-1996-4109-861a-d8f5d3177d5b] Reconfigured VM instance instance-00000075 to attach disk [datastore1] 7e32acbd-1996-4109-861a-d8f5d3177d5b/7e32acbd-1996-4109-861a-d8f5d3177d5b.vmdk or device None with type sparse {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1200.740755] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3c8c7b6b-b09c-4f4f-b0f3-bec552c743d0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.749849] env[63531]: DEBUG oslo_vmware.api [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Waiting for the task: (returnval){ [ 1200.749849] env[63531]: value = "task-1118774" [ 1200.749849] env[63531]: _type = "Task" [ 1200.749849] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1200.758041] env[63531]: DEBUG oslo_vmware.api [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Task: {'id': task-1118774, 'name': Rename_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.023362] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118773, 'name': CreateVM_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.143770] env[63531]: DEBUG nova.network.neutron [req-f5689a88-bac4-48ef-bf1b-ec6bf6dff669 req-a5fa2671-f2ea-4da3-9c4f-a59ad039d3e9 service nova] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Updated VIF entry in instance network info cache for port 9a3bf9ef-a105-4821-81f8-4891f7c3ae0a. {{(pid=63531) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1201.144159] env[63531]: DEBUG nova.network.neutron [req-f5689a88-bac4-48ef-bf1b-ec6bf6dff669 req-a5fa2671-f2ea-4da3-9c4f-a59ad039d3e9 service nova] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Updating instance_info_cache with network_info: [{"id": "9a3bf9ef-a105-4821-81f8-4891f7c3ae0a", "address": "fa:16:3e:dd:5e:bd", "network": {"id": "c4f4c994-36ab-42ab-94f8-78f29cbafa50", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-994507698-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "57e9bf4ee0f1463fa67b0f8deff30fdc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "767a3a48-41d4-4a0c-961d-0024837f63bd", "external-id": "nsx-vlan-transportzone-141", "segmentation_id": 141, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a3bf9ef-a1", "ovs_interfaceid": "9a3bf9ef-a105-4821-81f8-4891f7c3ae0a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1201.259871] env[63531]: DEBUG oslo_vmware.api [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Task: {'id': task-1118774, 'name': Rename_Task, 'duration_secs': 0.150168} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1201.260234] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] [instance: 7e32acbd-1996-4109-861a-d8f5d3177d5b] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1201.260344] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-44b714f5-dd02-465d-9e67-af5400ee1109 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.266335] env[63531]: DEBUG oslo_vmware.api [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Waiting for the task: (returnval){ [ 1201.266335] env[63531]: value = "task-1118775" [ 1201.266335] env[63531]: _type = "Task" [ 1201.266335] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1201.273238] env[63531]: DEBUG oslo_vmware.api [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Task: {'id': task-1118775, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.524908] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118773, 'name': CreateVM_Task} progress is 99%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.646808] env[63531]: DEBUG oslo_concurrency.lockutils [req-f5689a88-bac4-48ef-bf1b-ec6bf6dff669 req-a5fa2671-f2ea-4da3-9c4f-a59ad039d3e9 service nova] Releasing lock "refresh_cache-3794f44c-8f6d-4bab-b9e0-9f6c819e1a97" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1201.775962] env[63531]: DEBUG oslo_vmware.api [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Task: {'id': task-1118775, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1202.024449] env[63531]: DEBUG oslo_vmware.api [-] Task: {'id': task-1118773, 'name': CreateVM_Task, 'duration_secs': 1.347576} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1202.024622] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Created VM on the ESX host {{(pid=63531) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1202.025366] env[63531]: DEBUG oslo_concurrency.lockutils [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c6bead00-8690-442e-a30b-eb4e618e7b4b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1202.025585] env[63531]: DEBUG oslo_concurrency.lockutils [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c6bead00-8690-442e-a30b-eb4e618e7b4b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1202.026257] env[63531]: DEBUG oslo_concurrency.lockutils [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c6bead00-8690-442e-a30b-eb4e618e7b4b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1202.026257] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0f0da0ef-024b-497c-8e17-cbd410aa8ec0 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.030545] env[63531]: DEBUG oslo_vmware.api [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 1202.030545] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]525e68d7-a2c5-a375-1bbe-2445a31035ef" [ 1202.030545] env[63531]: _type = "Task" [ 1202.030545] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1202.038254] env[63531]: DEBUG oslo_vmware.api [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]525e68d7-a2c5-a375-1bbe-2445a31035ef, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1202.277943] env[63531]: DEBUG oslo_vmware.api [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Task: {'id': task-1118775, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1202.540742] env[63531]: DEBUG oslo_concurrency.lockutils [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c6bead00-8690-442e-a30b-eb4e618e7b4b" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1202.540962] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Processing image c6bead00-8690-442e-a30b-eb4e618e7b4b {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1202.541224] env[63531]: DEBUG oslo_concurrency.lockutils [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c6bead00-8690-442e-a30b-eb4e618e7b4b/c6bead00-8690-442e-a30b-eb4e618e7b4b.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1202.541378] env[63531]: DEBUG oslo_concurrency.lockutils [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c6bead00-8690-442e-a30b-eb4e618e7b4b/c6bead00-8690-442e-a30b-eb4e618e7b4b.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1202.541588] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1202.541839] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4ed12251-e2ef-47ea-bb7a-4d22e915b916 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.549918] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1202.550104] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63531) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1202.550737] env[63531]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-340d45c3-514b-4d72-860e-f4addeca8bbe {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.555116] env[63531]: DEBUG oslo_vmware.api [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 1202.555116] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]525b4237-bf1a-1df8-004c-42ad59dea432" [ 1202.555116] env[63531]: _type = "Task" [ 1202.555116] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1202.561872] env[63531]: DEBUG oslo_vmware.api [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': session[52de0a3e-63d7-580b-6611-524accfb0aed]525b4237-bf1a-1df8-004c-42ad59dea432, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1202.781308] env[63531]: DEBUG oslo_vmware.api [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Task: {'id': task-1118775, 'name': PowerOnVM_Task, 'duration_secs': 1.027777} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1202.781748] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] [instance: 7e32acbd-1996-4109-861a-d8f5d3177d5b] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1202.782106] env[63531]: INFO nova.compute.manager [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] [instance: 7e32acbd-1996-4109-861a-d8f5d3177d5b] Took 7.81 seconds to spawn the instance on the hypervisor. [ 1202.782444] env[63531]: DEBUG nova.compute.manager [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] [instance: 7e32acbd-1996-4109-861a-d8f5d3177d5b] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1202.783613] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcbffee6-0c64-4cff-be61-d6299c5c7efa {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.065773] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Preparing fetch location {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1203.066059] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Fetch image to [datastore1] OSTACK_IMG_81614cd5-4d5d-4f77-9b7b-cbd04319febe/OSTACK_IMG_81614cd5-4d5d-4f77-9b7b-cbd04319febe.vmdk {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1203.066253] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Downloading stream optimized image c6bead00-8690-442e-a30b-eb4e618e7b4b to [datastore1] OSTACK_IMG_81614cd5-4d5d-4f77-9b7b-cbd04319febe/OSTACK_IMG_81614cd5-4d5d-4f77-9b7b-cbd04319febe.vmdk on the data store datastore1 as vApp {{(pid=63531) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1203.066422] env[63531]: DEBUG nova.virt.vmwareapi.images [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Downloading image file data c6bead00-8690-442e-a30b-eb4e618e7b4b to the ESX as VM named 'OSTACK_IMG_81614cd5-4d5d-4f77-9b7b-cbd04319febe' {{(pid=63531) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1203.136256] env[63531]: DEBUG oslo_vmware.rw_handles [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1203.136256] env[63531]: value = "resgroup-9" [ 1203.136256] env[63531]: _type = "ResourcePool" [ 1203.136256] env[63531]: }. {{(pid=63531) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1203.136567] env[63531]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-87a3e667-fcfe-4890-8cc7-a4e5da02c109 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.157940] env[63531]: DEBUG oslo_vmware.rw_handles [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lease: (returnval){ [ 1203.157940] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52206bed-eecb-6b95-58f1-38d4ff2573cd" [ 1203.157940] env[63531]: _type = "HttpNfcLease" [ 1203.157940] env[63531]: } obtained for vApp import into resource pool (val){ [ 1203.157940] env[63531]: value = "resgroup-9" [ 1203.157940] env[63531]: _type = "ResourcePool" [ 1203.157940] env[63531]: }. {{(pid=63531) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1203.158269] env[63531]: DEBUG oslo_vmware.api [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the lease: (returnval){ [ 1203.158269] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52206bed-eecb-6b95-58f1-38d4ff2573cd" [ 1203.158269] env[63531]: _type = "HttpNfcLease" [ 1203.158269] env[63531]: } to be ready. {{(pid=63531) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1203.164750] env[63531]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1203.164750] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52206bed-eecb-6b95-58f1-38d4ff2573cd" [ 1203.164750] env[63531]: _type = "HttpNfcLease" [ 1203.164750] env[63531]: } is initializing. {{(pid=63531) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1203.304981] env[63531]: INFO nova.compute.manager [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] [instance: 7e32acbd-1996-4109-861a-d8f5d3177d5b] Took 14.59 seconds to build instance. [ 1203.666397] env[63531]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1203.666397] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52206bed-eecb-6b95-58f1-38d4ff2573cd" [ 1203.666397] env[63531]: _type = "HttpNfcLease" [ 1203.666397] env[63531]: } is ready. {{(pid=63531) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1203.666677] env[63531]: DEBUG oslo_vmware.rw_handles [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1203.666677] env[63531]: value = "session[52de0a3e-63d7-580b-6611-524accfb0aed]52206bed-eecb-6b95-58f1-38d4ff2573cd" [ 1203.666677] env[63531]: _type = "HttpNfcLease" [ 1203.666677] env[63531]: }. {{(pid=63531) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1203.667404] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83ed6b98-a3f2-4f11-ae02-8181265760fd {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.674455] env[63531]: DEBUG oslo_vmware.rw_handles [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/523a86f3-807e-1340-e6a2-4fd96b908218/disk-0.vmdk from lease info. {{(pid=63531) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1203.674639] env[63531]: DEBUG oslo_vmware.rw_handles [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Creating HTTP connection to write to file with size = 31667712 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/523a86f3-807e-1340-e6a2-4fd96b908218/disk-0.vmdk. {{(pid=63531) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1203.738650] env[63531]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-8d12326f-ce21-44c6-8bde-577d4191391a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.807775] env[63531]: DEBUG oslo_concurrency.lockutils [None req-536694ab-32de-4bbc-b8d1-886d50c97274 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Lock "7e32acbd-1996-4109-861a-d8f5d3177d5b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.101s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1204.334126] env[63531]: DEBUG oslo_concurrency.lockutils [None req-96eb5a78-8699-4444-a94d-d79663b747a8 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Acquiring lock "7e32acbd-1996-4109-861a-d8f5d3177d5b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1204.334532] env[63531]: DEBUG oslo_concurrency.lockutils [None req-96eb5a78-8699-4444-a94d-d79663b747a8 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Lock "7e32acbd-1996-4109-861a-d8f5d3177d5b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1204.334716] env[63531]: DEBUG oslo_concurrency.lockutils [None req-96eb5a78-8699-4444-a94d-d79663b747a8 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Acquiring lock "7e32acbd-1996-4109-861a-d8f5d3177d5b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1204.334962] env[63531]: DEBUG oslo_concurrency.lockutils [None req-96eb5a78-8699-4444-a94d-d79663b747a8 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Lock "7e32acbd-1996-4109-861a-d8f5d3177d5b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1204.335159] env[63531]: DEBUG oslo_concurrency.lockutils [None req-96eb5a78-8699-4444-a94d-d79663b747a8 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Lock "7e32acbd-1996-4109-861a-d8f5d3177d5b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1204.339218] env[63531]: INFO nova.compute.manager [None req-96eb5a78-8699-4444-a94d-d79663b747a8 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] [instance: 7e32acbd-1996-4109-861a-d8f5d3177d5b] Terminating instance [ 1204.341214] env[63531]: DEBUG nova.compute.manager [None req-96eb5a78-8699-4444-a94d-d79663b747a8 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] [instance: 7e32acbd-1996-4109-861a-d8f5d3177d5b] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1204.341457] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-96eb5a78-8699-4444-a94d-d79663b747a8 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] [instance: 7e32acbd-1996-4109-861a-d8f5d3177d5b] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1204.342481] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f917969-a36d-4ab5-90fe-c79eaa267ff9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.351018] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-96eb5a78-8699-4444-a94d-d79663b747a8 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] [instance: 7e32acbd-1996-4109-861a-d8f5d3177d5b] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1204.351260] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7bcf02ca-1701-4b08-b9ba-20f6496eaea7 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.358465] env[63531]: DEBUG oslo_vmware.api [None req-96eb5a78-8699-4444-a94d-d79663b747a8 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Waiting for the task: (returnval){ [ 1204.358465] env[63531]: value = "task-1118777" [ 1204.358465] env[63531]: _type = "Task" [ 1204.358465] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1204.367456] env[63531]: DEBUG oslo_vmware.api [None req-96eb5a78-8699-4444-a94d-d79663b747a8 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Task: {'id': task-1118777, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1204.848971] env[63531]: DEBUG oslo_vmware.rw_handles [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Completed reading data from the image iterator. {{(pid=63531) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1204.849338] env[63531]: DEBUG oslo_vmware.rw_handles [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/523a86f3-807e-1340-e6a2-4fd96b908218/disk-0.vmdk. {{(pid=63531) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1204.850339] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6adec19b-e0c2-4c79-ad8a-211667067f55 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.857499] env[63531]: DEBUG oslo_vmware.rw_handles [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/523a86f3-807e-1340-e6a2-4fd96b908218/disk-0.vmdk is in state: ready. {{(pid=63531) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1204.857769] env[63531]: DEBUG oslo_vmware.rw_handles [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/523a86f3-807e-1340-e6a2-4fd96b908218/disk-0.vmdk. {{(pid=63531) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1204.858096] env[63531]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-d2c840de-1f6a-49e2-9141-661fb2803a75 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.867601] env[63531]: DEBUG oslo_vmware.api [None req-96eb5a78-8699-4444-a94d-d79663b747a8 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Task: {'id': task-1118777, 'name': PowerOffVM_Task, 'duration_secs': 0.232487} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1204.867872] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-96eb5a78-8699-4444-a94d-d79663b747a8 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] [instance: 7e32acbd-1996-4109-861a-d8f5d3177d5b] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1204.868077] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-96eb5a78-8699-4444-a94d-d79663b747a8 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] [instance: 7e32acbd-1996-4109-861a-d8f5d3177d5b] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1204.868335] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-489f3e10-beee-4865-af6d-a08a8224a6d3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.933897] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-96eb5a78-8699-4444-a94d-d79663b747a8 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] [instance: 7e32acbd-1996-4109-861a-d8f5d3177d5b] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1204.934157] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-96eb5a78-8699-4444-a94d-d79663b747a8 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] [instance: 7e32acbd-1996-4109-861a-d8f5d3177d5b] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1204.934347] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-96eb5a78-8699-4444-a94d-d79663b747a8 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Deleting the datastore file [datastore1] 7e32acbd-1996-4109-861a-d8f5d3177d5b {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1204.934612] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9aceef4c-ad4d-44fe-95f2-bdfff152a3ae {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.940421] env[63531]: DEBUG oslo_vmware.api [None req-96eb5a78-8699-4444-a94d-d79663b747a8 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Waiting for the task: (returnval){ [ 1204.940421] env[63531]: value = "task-1118779" [ 1204.940421] env[63531]: _type = "Task" [ 1204.940421] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1204.947587] env[63531]: DEBUG oslo_vmware.api [None req-96eb5a78-8699-4444-a94d-d79663b747a8 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Task: {'id': task-1118779, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.065483] env[63531]: DEBUG oslo_vmware.rw_handles [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/523a86f3-807e-1340-e6a2-4fd96b908218/disk-0.vmdk. {{(pid=63531) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1205.065878] env[63531]: INFO nova.virt.vmwareapi.images [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Downloaded image file data c6bead00-8690-442e-a30b-eb4e618e7b4b [ 1205.066480] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a126eb8-90bf-4242-9ec3-db513655939a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.082165] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d6937f51-5a2f-4ab3-853c-539b3c211131 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.116161] env[63531]: INFO nova.virt.vmwareapi.images [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] The imported VM was unregistered [ 1205.118579] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Caching image {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1205.118855] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Creating directory with path [datastore1] devstack-image-cache_base/c6bead00-8690-442e-a30b-eb4e618e7b4b {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1205.119253] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8e19db79-ec4f-4261-8a8e-7bfc5de3097b {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.129175] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Created directory with path [datastore1] devstack-image-cache_base/c6bead00-8690-442e-a30b-eb4e618e7b4b {{(pid=63531) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1205.129368] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_81614cd5-4d5d-4f77-9b7b-cbd04319febe/OSTACK_IMG_81614cd5-4d5d-4f77-9b7b-cbd04319febe.vmdk to [datastore1] devstack-image-cache_base/c6bead00-8690-442e-a30b-eb4e618e7b4b/c6bead00-8690-442e-a30b-eb4e618e7b4b.vmdk. {{(pid=63531) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1205.129601] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-8c9081b6-487c-4d2b-8972-0751aee4959e {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.135500] env[63531]: DEBUG oslo_vmware.api [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 1205.135500] env[63531]: value = "task-1118781" [ 1205.135500] env[63531]: _type = "Task" [ 1205.135500] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1205.142203] env[63531]: DEBUG oslo_vmware.api [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118781, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.450144] env[63531]: DEBUG oslo_vmware.api [None req-96eb5a78-8699-4444-a94d-d79663b747a8 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Task: {'id': task-1118779, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.175314} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1205.450511] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-96eb5a78-8699-4444-a94d-d79663b747a8 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1205.450653] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-96eb5a78-8699-4444-a94d-d79663b747a8 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] [instance: 7e32acbd-1996-4109-861a-d8f5d3177d5b] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1205.450774] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-96eb5a78-8699-4444-a94d-d79663b747a8 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] [instance: 7e32acbd-1996-4109-861a-d8f5d3177d5b] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1205.450952] env[63531]: INFO nova.compute.manager [None req-96eb5a78-8699-4444-a94d-d79663b747a8 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] [instance: 7e32acbd-1996-4109-861a-d8f5d3177d5b] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1205.451215] env[63531]: DEBUG oslo.service.loopingcall [None req-96eb5a78-8699-4444-a94d-d79663b747a8 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1205.451417] env[63531]: DEBUG nova.compute.manager [-] [instance: 7e32acbd-1996-4109-861a-d8f5d3177d5b] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1205.451511] env[63531]: DEBUG nova.network.neutron [-] [instance: 7e32acbd-1996-4109-861a-d8f5d3177d5b] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1205.645956] env[63531]: DEBUG oslo_vmware.api [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118781, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.707116] env[63531]: DEBUG nova.compute.manager [req-5653f9d2-6741-4bf4-9de6-0ab9c25041e7 req-f27e95a5-f99f-4944-b79c-374ca9604d06 service nova] [instance: 7e32acbd-1996-4109-861a-d8f5d3177d5b] Received event network-vif-deleted-06ef1d52-c054-4224-9227-402b6a036abc {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1205.707512] env[63531]: INFO nova.compute.manager [req-5653f9d2-6741-4bf4-9de6-0ab9c25041e7 req-f27e95a5-f99f-4944-b79c-374ca9604d06 service nova] [instance: 7e32acbd-1996-4109-861a-d8f5d3177d5b] Neutron deleted interface 06ef1d52-c054-4224-9227-402b6a036abc; detaching it from the instance and deleting it from the info cache [ 1205.707686] env[63531]: DEBUG nova.network.neutron [req-5653f9d2-6741-4bf4-9de6-0ab9c25041e7 req-f27e95a5-f99f-4944-b79c-374ca9604d06 service nova] [instance: 7e32acbd-1996-4109-861a-d8f5d3177d5b] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1206.148399] env[63531]: DEBUG oslo_vmware.api [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118781, 'name': MoveVirtualDisk_Task} progress is 46%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.182035] env[63531]: DEBUG nova.network.neutron [-] [instance: 7e32acbd-1996-4109-861a-d8f5d3177d5b] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1206.210251] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0ce36532-3423-4a9a-978c-8fae7805d91f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.220756] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3106d231-828a-493b-a767-0041ecf3b0cc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.244978] env[63531]: DEBUG nova.compute.manager [req-5653f9d2-6741-4bf4-9de6-0ab9c25041e7 req-f27e95a5-f99f-4944-b79c-374ca9604d06 service nova] [instance: 7e32acbd-1996-4109-861a-d8f5d3177d5b] Detach interface failed, port_id=06ef1d52-c054-4224-9227-402b6a036abc, reason: Instance 7e32acbd-1996-4109-861a-d8f5d3177d5b could not be found. {{(pid=63531) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1206.648333] env[63531]: DEBUG oslo_vmware.api [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118781, 'name': MoveVirtualDisk_Task} progress is 66%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.685045] env[63531]: INFO nova.compute.manager [-] [instance: 7e32acbd-1996-4109-861a-d8f5d3177d5b] Took 1.23 seconds to deallocate network for instance. [ 1207.149041] env[63531]: DEBUG oslo_vmware.api [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118781, 'name': MoveVirtualDisk_Task} progress is 88%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1207.192486] env[63531]: DEBUG oslo_concurrency.lockutils [None req-96eb5a78-8699-4444-a94d-d79663b747a8 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1207.192770] env[63531]: DEBUG oslo_concurrency.lockutils [None req-96eb5a78-8699-4444-a94d-d79663b747a8 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1207.193034] env[63531]: DEBUG nova.objects.instance [None req-96eb5a78-8699-4444-a94d-d79663b747a8 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Lazy-loading 'resources' on Instance uuid 7e32acbd-1996-4109-861a-d8f5d3177d5b {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1207.649690] env[63531]: DEBUG oslo_vmware.api [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118781, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.271033} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1207.650079] env[63531]: INFO nova.virt.vmwareapi.ds_util [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_81614cd5-4d5d-4f77-9b7b-cbd04319febe/OSTACK_IMG_81614cd5-4d5d-4f77-9b7b-cbd04319febe.vmdk to [datastore1] devstack-image-cache_base/c6bead00-8690-442e-a30b-eb4e618e7b4b/c6bead00-8690-442e-a30b-eb4e618e7b4b.vmdk. [ 1207.650153] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Cleaning up location [datastore1] OSTACK_IMG_81614cd5-4d5d-4f77-9b7b-cbd04319febe {{(pid=63531) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1207.650320] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_81614cd5-4d5d-4f77-9b7b-cbd04319febe {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1207.650563] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fc6e7fba-d757-4cf3-9537-58ad903882f2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.656753] env[63531]: DEBUG oslo_vmware.api [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 1207.656753] env[63531]: value = "task-1118782" [ 1207.656753] env[63531]: _type = "Task" [ 1207.656753] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1207.664013] env[63531]: DEBUG oslo_vmware.api [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118782, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1207.736185] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24bcc020-ff89-4123-996a-d8a53c25dcfc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.742633] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-164656f4-26e8-4067-9bcb-4575be4b25ff {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.771576] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cca13c1b-4a3e-4d3f-beb1-2bf56130e520 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.778069] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae576d90-d167-4292-8c8f-b243ca652c42 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.790439] env[63531]: DEBUG nova.compute.provider_tree [None req-96eb5a78-8699-4444-a94d-d79663b747a8 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1208.167305] env[63531]: DEBUG oslo_vmware.api [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118782, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.038637} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1208.167511] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1208.167793] env[63531]: DEBUG oslo_concurrency.lockutils [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c6bead00-8690-442e-a30b-eb4e618e7b4b/c6bead00-8690-442e-a30b-eb4e618e7b4b.vmdk" {{(pid=63531) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1208.168122] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c6bead00-8690-442e-a30b-eb4e618e7b4b/c6bead00-8690-442e-a30b-eb4e618e7b4b.vmdk to [datastore1] 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97/3794f44c-8f6d-4bab-b9e0-9f6c819e1a97.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1208.168404] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a8357aca-4662-4fe7-bac6-b6b505a07fe3 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.175234] env[63531]: DEBUG oslo_vmware.api [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 1208.175234] env[63531]: value = "task-1118783" [ 1208.175234] env[63531]: _type = "Task" [ 1208.175234] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1208.182234] env[63531]: DEBUG oslo_vmware.api [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118783, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1208.294578] env[63531]: DEBUG nova.scheduler.client.report [None req-96eb5a78-8699-4444-a94d-d79663b747a8 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1208.685637] env[63531]: DEBUG oslo_vmware.api [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118783, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1208.799950] env[63531]: DEBUG oslo_concurrency.lockutils [None req-96eb5a78-8699-4444-a94d-d79663b747a8 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.607s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1208.819060] env[63531]: INFO nova.scheduler.client.report [None req-96eb5a78-8699-4444-a94d-d79663b747a8 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Deleted allocations for instance 7e32acbd-1996-4109-861a-d8f5d3177d5b [ 1209.185570] env[63531]: DEBUG oslo_vmware.api [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118783, 'name': CopyVirtualDisk_Task} progress is 46%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1209.327816] env[63531]: DEBUG oslo_concurrency.lockutils [None req-96eb5a78-8699-4444-a94d-d79663b747a8 tempest-ServerTagsTestJSON-2122360734 tempest-ServerTagsTestJSON-2122360734-project-member] Lock "7e32acbd-1996-4109-861a-d8f5d3177d5b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.993s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1209.687446] env[63531]: DEBUG oslo_vmware.api [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118783, 'name': CopyVirtualDisk_Task} progress is 69%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.188950] env[63531]: DEBUG oslo_vmware.api [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118783, 'name': CopyVirtualDisk_Task} progress is 85%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.689212] env[63531]: DEBUG oslo_vmware.api [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118783, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1211.189940] env[63531]: DEBUG oslo_vmware.api [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118783, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.525903} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1211.190254] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c6bead00-8690-442e-a30b-eb4e618e7b4b/c6bead00-8690-442e-a30b-eb4e618e7b4b.vmdk to [datastore1] 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97/3794f44c-8f6d-4bab-b9e0-9f6c819e1a97.vmdk {{(pid=63531) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1211.191078] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de7d316f-3dd2-4633-acd0-87132d70f93a {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.213269] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Reconfiguring VM instance instance-00000072 to attach disk [datastore1] 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97/3794f44c-8f6d-4bab-b9e0-9f6c819e1a97.vmdk or device None with type streamOptimized {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1211.213593] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9e867ddf-059d-4395-91dd-058b16193809 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.233096] env[63531]: DEBUG oslo_vmware.api [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 1211.233096] env[63531]: value = "task-1118784" [ 1211.233096] env[63531]: _type = "Task" [ 1211.233096] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1211.240924] env[63531]: DEBUG oslo_vmware.api [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118784, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1211.743781] env[63531]: DEBUG oslo_vmware.api [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118784, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1212.243323] env[63531]: DEBUG oslo_vmware.api [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118784, 'name': ReconfigVM_Task, 'duration_secs': 0.588544} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1212.243610] env[63531]: DEBUG nova.virt.vmwareapi.volumeops [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Reconfigured VM instance instance-00000072 to attach disk [datastore1] 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97/3794f44c-8f6d-4bab-b9e0-9f6c819e1a97.vmdk or device None with type streamOptimized {{(pid=63531) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1212.244268] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0d37282b-ff29-44f7-9722-40553ee06af9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.251030] env[63531]: DEBUG oslo_vmware.api [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 1212.251030] env[63531]: value = "task-1118785" [ 1212.251030] env[63531]: _type = "Task" [ 1212.251030] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1212.258083] env[63531]: DEBUG oslo_vmware.api [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118785, 'name': Rename_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1212.760547] env[63531]: DEBUG oslo_vmware.api [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118785, 'name': Rename_Task, 'duration_secs': 0.152514} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1212.760915] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Powering on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1212.761093] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-062ce102-3173-4b24-af76-6fb87ade56d9 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.766950] env[63531]: DEBUG oslo_vmware.api [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 1212.766950] env[63531]: value = "task-1118786" [ 1212.766950] env[63531]: _type = "Task" [ 1212.766950] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1212.775249] env[63531]: DEBUG oslo_vmware.api [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118786, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1213.277078] env[63531]: DEBUG oslo_vmware.api [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118786, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1213.778315] env[63531]: DEBUG oslo_vmware.api [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118786, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.281240] env[63531]: DEBUG oslo_vmware.api [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118786, 'name': PowerOnVM_Task, 'duration_secs': 1.02001} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1214.281628] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Powered on the VM {{(pid=63531) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1214.390941] env[63531]: DEBUG nova.compute.manager [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Checking state {{(pid=63531) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1214.391979] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5889d54-cdb4-4158-9780-48d006ad16e2 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.909040] env[63531]: DEBUG oslo_concurrency.lockutils [None req-887422b4-1d23-451d-8c49-4ef19676ecf8 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "3794f44c-8f6d-4bab-b9e0-9f6c819e1a97" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 23.591s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1216.267055] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c7645f10-568d-4e4e-86c3-a8c62dd8c7e4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquiring lock "3794f44c-8f6d-4bab-b9e0-9f6c819e1a97" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1216.267055] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c7645f10-568d-4e4e-86c3-a8c62dd8c7e4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "3794f44c-8f6d-4bab-b9e0-9f6c819e1a97" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1216.267055] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c7645f10-568d-4e4e-86c3-a8c62dd8c7e4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquiring lock "3794f44c-8f6d-4bab-b9e0-9f6c819e1a97-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1216.267055] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c7645f10-568d-4e4e-86c3-a8c62dd8c7e4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "3794f44c-8f6d-4bab-b9e0-9f6c819e1a97-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1216.267055] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c7645f10-568d-4e4e-86c3-a8c62dd8c7e4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "3794f44c-8f6d-4bab-b9e0-9f6c819e1a97-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1216.269194] env[63531]: INFO nova.compute.manager [None req-c7645f10-568d-4e4e-86c3-a8c62dd8c7e4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Terminating instance [ 1216.270962] env[63531]: DEBUG nova.compute.manager [None req-c7645f10-568d-4e4e-86c3-a8c62dd8c7e4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Start destroying the instance on the hypervisor. {{(pid=63531) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1216.271180] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c7645f10-568d-4e4e-86c3-a8c62dd8c7e4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Destroying instance {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1216.272062] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-293edf4d-7e9d-477c-94b3-d733ced48c50 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.279954] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7645f10-568d-4e4e-86c3-a8c62dd8c7e4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Powering off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1216.280203] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ff50bcaa-c0da-4127-be66-760871559263 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.286552] env[63531]: DEBUG oslo_vmware.api [None req-c7645f10-568d-4e4e-86c3-a8c62dd8c7e4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 1216.286552] env[63531]: value = "task-1118787" [ 1216.286552] env[63531]: _type = "Task" [ 1216.286552] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1216.294044] env[63531]: DEBUG oslo_vmware.api [None req-c7645f10-568d-4e4e-86c3-a8c62dd8c7e4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118787, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1216.796382] env[63531]: DEBUG oslo_vmware.api [None req-c7645f10-568d-4e4e-86c3-a8c62dd8c7e4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118787, 'name': PowerOffVM_Task, 'duration_secs': 0.185253} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1216.796648] env[63531]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7645f10-568d-4e4e-86c3-a8c62dd8c7e4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Powered off the VM {{(pid=63531) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1216.796823] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c7645f10-568d-4e4e-86c3-a8c62dd8c7e4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Unregistering the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1216.797081] env[63531]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ec949979-0372-4662-92a6-614c9f596c44 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.861971] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c7645f10-568d-4e4e-86c3-a8c62dd8c7e4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Unregistered the VM {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1216.862213] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c7645f10-568d-4e4e-86c3-a8c62dd8c7e4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Deleting contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1216.862402] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7645f10-568d-4e4e-86c3-a8c62dd8c7e4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Deleting the datastore file [datastore1] 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97 {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1216.862657] env[63531]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a2a959e3-e948-4d67-a44c-dd82b9981921 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.868816] env[63531]: DEBUG oslo_vmware.api [None req-c7645f10-568d-4e4e-86c3-a8c62dd8c7e4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for the task: (returnval){ [ 1216.868816] env[63531]: value = "task-1118789" [ 1216.868816] env[63531]: _type = "Task" [ 1216.868816] env[63531]: } to complete. {{(pid=63531) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1216.875862] env[63531]: DEBUG oslo_vmware.api [None req-c7645f10-568d-4e4e-86c3-a8c62dd8c7e4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118789, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1217.379125] env[63531]: DEBUG oslo_vmware.api [None req-c7645f10-568d-4e4e-86c3-a8c62dd8c7e4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Task: {'id': task-1118789, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.126577} completed successfully. {{(pid=63531) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1217.379496] env[63531]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7645f10-568d-4e4e-86c3-a8c62dd8c7e4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Deleted the datastore file {{(pid=63531) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1217.379597] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c7645f10-568d-4e4e-86c3-a8c62dd8c7e4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Deleted contents of the VM from datastore datastore1 {{(pid=63531) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1217.379746] env[63531]: DEBUG nova.virt.vmwareapi.vmops [None req-c7645f10-568d-4e4e-86c3-a8c62dd8c7e4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Instance destroyed {{(pid=63531) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1217.379928] env[63531]: INFO nova.compute.manager [None req-c7645f10-568d-4e4e-86c3-a8c62dd8c7e4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1217.380194] env[63531]: DEBUG oslo.service.loopingcall [None req-c7645f10-568d-4e4e-86c3-a8c62dd8c7e4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63531) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1217.380381] env[63531]: DEBUG nova.compute.manager [-] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Deallocating network for instance {{(pid=63531) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1217.380475] env[63531]: DEBUG nova.network.neutron [-] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] deallocate_for_instance() {{(pid=63531) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1217.797441] env[63531]: DEBUG nova.compute.manager [req-79d92d6b-66af-4166-85d8-e109220685b8 req-1066926d-50f4-458e-8d25-a5b6ab2b1cd9 service nova] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Received event network-vif-deleted-9a3bf9ef-a105-4821-81f8-4891f7c3ae0a {{(pid=63531) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1217.797690] env[63531]: INFO nova.compute.manager [req-79d92d6b-66af-4166-85d8-e109220685b8 req-1066926d-50f4-458e-8d25-a5b6ab2b1cd9 service nova] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Neutron deleted interface 9a3bf9ef-a105-4821-81f8-4891f7c3ae0a; detaching it from the instance and deleting it from the info cache [ 1217.797967] env[63531]: DEBUG nova.network.neutron [req-79d92d6b-66af-4166-85d8-e109220685b8 req-1066926d-50f4-458e-8d25-a5b6ab2b1cd9 service nova] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1218.265409] env[63531]: DEBUG nova.network.neutron [-] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Updating instance_info_cache with network_info: [] {{(pid=63531) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1218.301123] env[63531]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-66a79269-2009-4553-ba83-e41decede16f {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.312385] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0234583-5d7e-49d6-a303-6f192207496c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.335324] env[63531]: DEBUG nova.compute.manager [req-79d92d6b-66af-4166-85d8-e109220685b8 req-1066926d-50f4-458e-8d25-a5b6ab2b1cd9 service nova] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Detach interface failed, port_id=9a3bf9ef-a105-4821-81f8-4891f7c3ae0a, reason: Instance 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97 could not be found. {{(pid=63531) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1218.768206] env[63531]: INFO nova.compute.manager [-] [instance: 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97] Took 1.39 seconds to deallocate network for instance. [ 1219.274973] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c7645f10-568d-4e4e-86c3-a8c62dd8c7e4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1219.275267] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c7645f10-568d-4e4e-86c3-a8c62dd8c7e4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1219.275499] env[63531]: DEBUG nova.objects.instance [None req-c7645f10-568d-4e4e-86c3-a8c62dd8c7e4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lazy-loading 'resources' on Instance uuid 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97 {{(pid=63531) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1219.810893] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11d2c70b-1b0d-4a18-841a-7dfc87a5c044 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.817765] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-530fccd8-2397-40c5-92c1-98586dbfb926 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.846565] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26265bed-5bcd-4310-ac46-e031db258558 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.853340] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0604c774-a2cd-4695-bfc9-41520195a653 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.867182] env[63531]: DEBUG nova.compute.provider_tree [None req-c7645f10-568d-4e4e-86c3-a8c62dd8c7e4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1220.370619] env[63531]: DEBUG nova.scheduler.client.report [None req-c7645f10-568d-4e4e-86c3-a8c62dd8c7e4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1220.876133] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c7645f10-568d-4e4e-86c3-a8c62dd8c7e4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.601s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1220.895855] env[63531]: INFO nova.scheduler.client.report [None req-c7645f10-568d-4e4e-86c3-a8c62dd8c7e4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Deleted allocations for instance 3794f44c-8f6d-4bab-b9e0-9f6c819e1a97 [ 1221.403864] env[63531]: DEBUG oslo_concurrency.lockutils [None req-c7645f10-568d-4e4e-86c3-a8c62dd8c7e4 tempest-ServerActionsTestOtherB-1713416238 tempest-ServerActionsTestOtherB-1713416238-project-member] Lock "3794f44c-8f6d-4bab-b9e0-9f6c819e1a97" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.137s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1221.778131] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1222.777758] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1224.777596] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1224.777881] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Starting heal instance info cache {{(pid=63531) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1224.777881] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Rebuilding the list of instances to heal {{(pid=63531) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1225.280372] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Didn't find any instances for network info cache update. {{(pid=63531) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1225.280631] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1225.280774] env[63531]: DEBUG nova.compute.manager [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63531) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1226.276482] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1226.777764] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1227.778126] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1227.778485] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1228.778071] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager.update_available_resource {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1229.282021] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1229.282383] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1229.282522] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1229.282682] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63531) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1229.283600] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-304b5bd8-1e69-4248-8bf7-8c42c87ab552 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.292326] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e63d25d9-b592-42b8-a30f-09c40d28db80 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.306265] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d221539-e08e-4ed2-97c0-5cc6a23bacc4 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.312186] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9623a5af-4bc3-4c37-9042-e82a4b83162c {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.340686] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181461MB free_disk=170GB free_vcpus=48 pci_devices=None {{(pid=63531) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1229.340827] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1229.341017] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1230.360530] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=63531) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1230.360812] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=63531) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1230.373490] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fd9cd6c-0ceb-44d0-8a48-d6fde4aeb077 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.381840] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0385d63-033e-4d0f-b202-53247cb80bdc {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.410337] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3646d90-3a17-462d-a705-64445d9badc1 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.417200] env[63531]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08aa7736-d1d1-4c5a-8259-1b1144bf3767 {{(pid=63531) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.429579] env[63531]: DEBUG nova.compute.provider_tree [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Inventory has not changed in ProviderTree for provider: 9996a314-a62e-4282-8252-f387e7e3fabb {{(pid=63531) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1230.933128] env[63531]: DEBUG nova.scheduler.client.report [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Inventory has not changed for provider 9996a314-a62e-4282-8252-f387e7e3fabb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63531) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1231.438807] env[63531]: DEBUG nova.compute.resource_tracker [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63531) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1231.439128] env[63531]: DEBUG oslo_concurrency.lockutils [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.098s {{(pid=63531) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1232.434727] env[63531]: DEBUG oslo_service.periodic_task [None req-6d574f02-f163-4153-bf08-2ed0120902cf None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63531) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}}